00:00:00.001 Started by upstream project "autotest-per-patch" build number 132395 00:00:00.001 originally caused by: 00:00:00.001 Started by user sys_sgci 00:00:00.011 Checking out git https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool into /var/jenkins_home/workspace/raid-vg-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4 to read jbp/jenkins/jjb-config/jobs/autotest-downstream/autotest-vg.groovy 00:00:09.326 The recommended git tool is: git 00:00:09.327 using credential 00000000-0000-0000-0000-000000000002 00:00:09.331 > git rev-parse --resolve-git-dir /var/jenkins_home/workspace/raid-vg-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4/jbp/.git # timeout=10 00:00:09.343 Fetching changes from the remote Git repository 00:00:09.348 > git config remote.origin.url https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool # timeout=10 00:00:09.360 Using shallow fetch with depth 1 00:00:09.360 Fetching upstream changes from https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool 00:00:09.360 > git --version # timeout=10 00:00:09.373 > git --version # 'git version 2.39.2' 00:00:09.373 using GIT_ASKPASS to set credentials SPDKCI HTTPS Credentials 00:00:09.386 Setting http proxy: proxy-dmz.intel.com:911 00:00:09.386 > git fetch --tags --force --progress --depth=1 -- https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool refs/heads/master # timeout=5 00:00:15.119 > git rev-parse origin/FETCH_HEAD^{commit} # timeout=10 00:00:15.131 > git rev-parse FETCH_HEAD^{commit} # timeout=10 00:00:15.142 Checking out Revision db4637e8b949f278f369ec13f70585206ccd9507 (FETCH_HEAD) 00:00:15.142 > git config core.sparsecheckout # timeout=10 00:00:15.154 > git read-tree -mu HEAD # timeout=10 00:00:15.171 > git checkout -f db4637e8b949f278f369ec13f70585206ccd9507 # timeout=5 00:00:15.195 Commit message: "jenkins/jjb-config: Add missing SPDK_TEST_NVME_INTERRUPT flag" 00:00:15.195 > git rev-list --no-walk db4637e8b949f278f369ec13f70585206ccd9507 # timeout=10 00:00:15.296 [Pipeline] Start of Pipeline 00:00:15.311 [Pipeline] library 00:00:15.313 Loading library shm_lib@master 00:00:15.313 Library shm_lib@master is cached. Copying from home. 00:00:15.329 [Pipeline] node 00:00:30.331 Still waiting to schedule task 00:00:30.331 Waiting for next available executor on ‘vagrant-vm-host’ 00:21:27.238 Running on VM-host-SM38 in /var/jenkins/workspace/raid-vg-autotest 00:21:27.240 [Pipeline] { 00:21:27.250 [Pipeline] catchError 00:21:27.252 [Pipeline] { 00:21:27.265 [Pipeline] wrap 00:21:27.274 [Pipeline] { 00:21:27.282 [Pipeline] stage 00:21:27.283 [Pipeline] { (Prologue) 00:21:27.302 [Pipeline] echo 00:21:27.303 Node: VM-host-SM38 00:21:27.310 [Pipeline] cleanWs 00:21:27.325 [WS-CLEANUP] Deleting project workspace... 00:21:27.325 [WS-CLEANUP] Deferred wipeout is used... 00:21:27.330 [WS-CLEANUP] done 00:21:27.583 [Pipeline] setCustomBuildProperty 00:21:27.683 [Pipeline] httpRequest 00:21:27.999 [Pipeline] echo 00:21:28.001 Sorcerer 10.211.164.20 is alive 00:21:28.014 [Pipeline] retry 00:21:28.016 [Pipeline] { 00:21:28.036 [Pipeline] httpRequest 00:21:28.041 HttpMethod: GET 00:21:28.042 URL: http://10.211.164.20/packages/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:21:28.042 Sending request to url: http://10.211.164.20/packages/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:21:28.043 Response Code: HTTP/1.1 200 OK 00:21:28.044 Success: Status code 200 is in the accepted range: 200,404 00:21:28.044 Saving response body to /var/jenkins/workspace/raid-vg-autotest/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:21:28.190 [Pipeline] } 00:21:28.212 [Pipeline] // retry 00:21:28.222 [Pipeline] sh 00:21:28.505 + tar --no-same-owner -xf jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:21:28.522 [Pipeline] httpRequest 00:21:28.831 [Pipeline] echo 00:21:28.833 Sorcerer 10.211.164.20 is alive 00:21:28.845 [Pipeline] retry 00:21:28.848 [Pipeline] { 00:21:28.864 [Pipeline] httpRequest 00:21:28.869 HttpMethod: GET 00:21:28.870 URL: http://10.211.164.20/packages/spdk_a361eb5e2807baab35986e5161b461bb8015fc19.tar.gz 00:21:28.870 Sending request to url: http://10.211.164.20/packages/spdk_a361eb5e2807baab35986e5161b461bb8015fc19.tar.gz 00:21:28.871 Response Code: HTTP/1.1 200 OK 00:21:28.871 Success: Status code 200 is in the accepted range: 200,404 00:21:28.872 Saving response body to /var/jenkins/workspace/raid-vg-autotest/spdk_a361eb5e2807baab35986e5161b461bb8015fc19.tar.gz 00:21:31.139 [Pipeline] } 00:21:31.159 [Pipeline] // retry 00:21:31.171 [Pipeline] sh 00:21:31.449 + tar --no-same-owner -xf spdk_a361eb5e2807baab35986e5161b461bb8015fc19.tar.gz 00:21:34.810 [Pipeline] sh 00:21:35.085 + git -C spdk log --oneline -n5 00:21:35.085 a361eb5e2 nvme_spec: Add SPDK_NVME_IO_FLAGS_PRCHK_MASK 00:21:35.085 4ab755590 bdev: Insert or overwrite metadata using bounce/accel buffer if NVMe PRACT is set 00:21:35.085 f40c2e7bb dif: Add spdk_dif_pi_format_get_pi_size() to use for NVMe PRACT 00:21:35.085 325a79ea3 bdev/malloc: Support accel sequence when DIF is enabled 00:21:35.085 0b4b4be7e bdev: Add spdk_bdev_io_hide_metadata() for bdev modules 00:21:35.102 [Pipeline] writeFile 00:21:35.117 [Pipeline] sh 00:21:35.394 + jbp/jenkins/jjb-config/jobs/scripts/autorun_quirks.sh 00:21:35.403 [Pipeline] sh 00:21:35.677 + cat autorun-spdk.conf 00:21:35.677 SPDK_RUN_FUNCTIONAL_TEST=1 00:21:35.677 SPDK_RUN_ASAN=1 00:21:35.677 SPDK_RUN_UBSAN=1 00:21:35.677 SPDK_TEST_RAID=1 00:21:35.677 SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:21:35.681 RUN_NIGHTLY=0 00:21:35.682 [Pipeline] } 00:21:35.694 [Pipeline] // stage 00:21:35.707 [Pipeline] stage 00:21:35.709 [Pipeline] { (Run VM) 00:21:35.720 [Pipeline] sh 00:21:36.066 + jbp/jenkins/jjb-config/jobs/scripts/prepare_nvme.sh 00:21:36.066 + echo 'Start stage prepare_nvme.sh' 00:21:36.066 Start stage prepare_nvme.sh 00:21:36.066 + [[ -n 0 ]] 00:21:36.066 + disk_prefix=ex0 00:21:36.066 + [[ -n /var/jenkins/workspace/raid-vg-autotest ]] 00:21:36.066 + [[ -e /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf ]] 00:21:36.066 + source /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf 00:21:36.066 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:21:36.066 ++ SPDK_RUN_ASAN=1 00:21:36.066 ++ SPDK_RUN_UBSAN=1 00:21:36.066 ++ SPDK_TEST_RAID=1 00:21:36.066 ++ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:21:36.066 ++ RUN_NIGHTLY=0 00:21:36.066 + cd /var/jenkins/workspace/raid-vg-autotest 00:21:36.066 + nvme_files=() 00:21:36.066 + declare -A nvme_files 00:21:36.066 + backend_dir=/var/lib/libvirt/images/backends 00:21:36.066 + nvme_files['nvme.img']=5G 00:21:36.066 + nvme_files['nvme-cmb.img']=5G 00:21:36.066 + nvme_files['nvme-multi0.img']=4G 00:21:36.066 + nvme_files['nvme-multi1.img']=4G 00:21:36.066 + nvme_files['nvme-multi2.img']=4G 00:21:36.066 + nvme_files['nvme-openstack.img']=8G 00:21:36.066 + nvme_files['nvme-zns.img']=5G 00:21:36.066 + (( SPDK_TEST_NVME_PMR == 1 )) 00:21:36.066 + (( SPDK_TEST_FTL == 1 )) 00:21:36.066 + (( SPDK_TEST_NVME_FDP == 1 )) 00:21:36.066 + [[ ! -d /var/lib/libvirt/images/backends ]] 00:21:36.066 + for nvme in "${!nvme_files[@]}" 00:21:36.066 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex0-nvme-multi2.img -s 4G 00:21:36.066 Formatting '/var/lib/libvirt/images/backends/ex0-nvme-multi2.img', fmt=raw size=4294967296 preallocation=falloc 00:21:36.066 + for nvme in "${!nvme_files[@]}" 00:21:36.066 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex0-nvme-cmb.img -s 5G 00:21:36.066 Formatting '/var/lib/libvirt/images/backends/ex0-nvme-cmb.img', fmt=raw size=5368709120 preallocation=falloc 00:21:36.066 + for nvme in "${!nvme_files[@]}" 00:21:36.066 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex0-nvme-openstack.img -s 8G 00:21:36.066 Formatting '/var/lib/libvirt/images/backends/ex0-nvme-openstack.img', fmt=raw size=8589934592 preallocation=falloc 00:21:36.066 + for nvme in "${!nvme_files[@]}" 00:21:36.066 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex0-nvme-zns.img -s 5G 00:21:36.066 Formatting '/var/lib/libvirt/images/backends/ex0-nvme-zns.img', fmt=raw size=5368709120 preallocation=falloc 00:21:36.066 + for nvme in "${!nvme_files[@]}" 00:21:36.066 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex0-nvme-multi1.img -s 4G 00:21:36.066 Formatting '/var/lib/libvirt/images/backends/ex0-nvme-multi1.img', fmt=raw size=4294967296 preallocation=falloc 00:21:36.066 + for nvme in "${!nvme_files[@]}" 00:21:36.066 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex0-nvme-multi0.img -s 4G 00:21:36.066 Formatting '/var/lib/libvirt/images/backends/ex0-nvme-multi0.img', fmt=raw size=4294967296 preallocation=falloc 00:21:36.066 + for nvme in "${!nvme_files[@]}" 00:21:36.066 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex0-nvme.img -s 5G 00:21:36.066 Formatting '/var/lib/libvirt/images/backends/ex0-nvme.img', fmt=raw size=5368709120 preallocation=falloc 00:21:36.066 ++ sudo grep -rl ex0-nvme.img /etc/libvirt/qemu 00:21:36.066 + echo 'End stage prepare_nvme.sh' 00:21:36.066 End stage prepare_nvme.sh 00:21:36.077 [Pipeline] sh 00:21:36.353 + DISTRO=fedora39 00:21:36.353 + CPUS=10 00:21:36.353 + RAM=12288 00:21:36.353 + jbp/jenkins/jjb-config/jobs/scripts/vagrant_create_vm.sh 00:21:36.353 Setup: -n 10 -s 12288 -x -p libvirt --qemu-emulator=/usr/local/qemu/vanilla-v8.0.0/bin/qemu-system-x86_64 --nic-model=e1000 -b /var/lib/libvirt/images/backends/ex0-nvme.img -b /var/lib/libvirt/images/backends/ex0-nvme-multi0.img,nvme,/var/lib/libvirt/images/backends/ex0-nvme-multi1.img:/var/lib/libvirt/images/backends/ex0-nvme-multi2.img -H -a -v -f fedora39 00:21:36.353 00:21:36.353 DIR=/var/jenkins/workspace/raid-vg-autotest/spdk/scripts/vagrant 00:21:36.353 SPDK_DIR=/var/jenkins/workspace/raid-vg-autotest/spdk 00:21:36.353 VAGRANT_TARGET=/var/jenkins/workspace/raid-vg-autotest 00:21:36.353 HELP=0 00:21:36.353 DRY_RUN=0 00:21:36.353 NVME_FILE=/var/lib/libvirt/images/backends/ex0-nvme.img,/var/lib/libvirt/images/backends/ex0-nvme-multi0.img, 00:21:36.353 NVME_DISKS_TYPE=nvme,nvme, 00:21:36.353 NVME_AUTO_CREATE=0 00:21:36.353 NVME_DISKS_NAMESPACES=,/var/lib/libvirt/images/backends/ex0-nvme-multi1.img:/var/lib/libvirt/images/backends/ex0-nvme-multi2.img, 00:21:36.353 NVME_CMB=,, 00:21:36.353 NVME_PMR=,, 00:21:36.353 NVME_ZNS=,, 00:21:36.353 NVME_MS=,, 00:21:36.353 NVME_FDP=,, 00:21:36.353 SPDK_VAGRANT_DISTRO=fedora39 00:21:36.353 SPDK_VAGRANT_VMCPU=10 00:21:36.353 SPDK_VAGRANT_VMRAM=12288 00:21:36.353 SPDK_VAGRANT_PROVIDER=libvirt 00:21:36.353 SPDK_VAGRANT_HTTP_PROXY= 00:21:36.353 SPDK_QEMU_EMULATOR=/usr/local/qemu/vanilla-v8.0.0/bin/qemu-system-x86_64 00:21:36.353 SPDK_OPENSTACK_NETWORK=0 00:21:36.353 VAGRANT_PACKAGE_BOX=0 00:21:36.353 VAGRANTFILE=/var/jenkins/workspace/raid-vg-autotest/spdk/scripts/vagrant/Vagrantfile 00:21:36.353 FORCE_DISTRO=true 00:21:36.353 VAGRANT_BOX_VERSION= 00:21:36.353 EXTRA_VAGRANTFILES= 00:21:36.353 NIC_MODEL=e1000 00:21:36.353 00:21:36.353 mkdir: created directory '/var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt' 00:21:36.353 /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt /var/jenkins/workspace/raid-vg-autotest 00:21:38.877 Bringing machine 'default' up with 'libvirt' provider... 00:21:39.135 ==> default: Creating image (snapshot of base box volume). 00:21:39.135 ==> default: Creating domain with the following settings... 00:21:39.135 ==> default: -- Name: fedora39-39-1.5-1721788873-2326_default_1732110105_9cbb1b724db4c2557612 00:21:39.135 ==> default: -- Domain type: kvm 00:21:39.135 ==> default: -- Cpus: 10 00:21:39.135 ==> default: -- Feature: acpi 00:21:39.135 ==> default: -- Feature: apic 00:21:39.135 ==> default: -- Feature: pae 00:21:39.135 ==> default: -- Memory: 12288M 00:21:39.135 ==> default: -- Memory Backing: hugepages: 00:21:39.135 ==> default: -- Management MAC: 00:21:39.135 ==> default: -- Loader: 00:21:39.135 ==> default: -- Nvram: 00:21:39.135 ==> default: -- Base box: spdk/fedora39 00:21:39.135 ==> default: -- Storage pool: default 00:21:39.135 ==> default: -- Image: /var/lib/libvirt/images/fedora39-39-1.5-1721788873-2326_default_1732110105_9cbb1b724db4c2557612.img (20G) 00:21:39.135 ==> default: -- Volume Cache: default 00:21:39.135 ==> default: -- Kernel: 00:21:39.135 ==> default: -- Initrd: 00:21:39.135 ==> default: -- Graphics Type: vnc 00:21:39.135 ==> default: -- Graphics Port: -1 00:21:39.135 ==> default: -- Graphics IP: 127.0.0.1 00:21:39.135 ==> default: -- Graphics Password: Not defined 00:21:39.135 ==> default: -- Video Type: cirrus 00:21:39.135 ==> default: -- Video VRAM: 9216 00:21:39.135 ==> default: -- Sound Type: 00:21:39.135 ==> default: -- Keymap: en-us 00:21:39.135 ==> default: -- TPM Path: 00:21:39.135 ==> default: -- INPUT: type=mouse, bus=ps2 00:21:39.135 ==> default: -- Command line args: 00:21:39.135 ==> default: -> value=-device, 00:21:39.135 ==> default: -> value=nvme,id=nvme-0,serial=12340,addr=0x10, 00:21:39.135 ==> default: -> value=-drive, 00:21:39.135 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex0-nvme.img,if=none,id=nvme-0-drive0, 00:21:39.135 ==> default: -> value=-device, 00:21:39.135 ==> default: -> value=nvme-ns,drive=nvme-0-drive0,bus=nvme-0,nsid=1,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:21:39.135 ==> default: -> value=-device, 00:21:39.136 ==> default: -> value=nvme,id=nvme-1,serial=12341,addr=0x11, 00:21:39.136 ==> default: -> value=-drive, 00:21:39.136 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex0-nvme-multi0.img,if=none,id=nvme-1-drive0, 00:21:39.136 ==> default: -> value=-device, 00:21:39.136 ==> default: -> value=nvme-ns,drive=nvme-1-drive0,bus=nvme-1,nsid=1,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:21:39.136 ==> default: -> value=-drive, 00:21:39.136 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex0-nvme-multi1.img,if=none,id=nvme-1-drive1, 00:21:39.136 ==> default: -> value=-device, 00:21:39.136 ==> default: -> value=nvme-ns,drive=nvme-1-drive1,bus=nvme-1,nsid=2,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:21:39.136 ==> default: -> value=-drive, 00:21:39.136 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex0-nvme-multi2.img,if=none,id=nvme-1-drive2, 00:21:39.136 ==> default: -> value=-device, 00:21:39.136 ==> default: -> value=nvme-ns,drive=nvme-1-drive2,bus=nvme-1,nsid=3,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:21:39.394 ==> default: Creating shared folders metadata... 00:21:39.394 ==> default: Starting domain. 00:21:40.327 ==> default: Waiting for domain to get an IP address... 00:21:58.465 ==> default: Waiting for SSH to become available... 00:21:58.465 ==> default: Configuring and enabling network interfaces... 00:22:00.362 default: SSH address: 192.168.121.104:22 00:22:00.362 default: SSH username: vagrant 00:22:00.362 default: SSH auth method: private key 00:22:02.301 ==> default: Rsyncing folder: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/spdk/ => /home/vagrant/spdk_repo/spdk 00:22:12.306 ==> default: Mounting SSHFS shared folder... 00:22:12.306 ==> default: Mounting folder via SSHFS: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/fedora39-libvirt/output => /home/vagrant/spdk_repo/output 00:22:12.306 ==> default: Checking Mount.. 00:22:13.290 ==> default: Folder Successfully Mounted! 00:22:13.290 00:22:13.290 SUCCESS! 00:22:13.290 00:22:13.290 cd to /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt and type "vagrant ssh" to use. 00:22:13.290 Use vagrant "suspend" and vagrant "resume" to stop and start. 00:22:13.290 Use vagrant "destroy" followed by "rm -rf /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt" to destroy all trace of vm. 00:22:13.290 00:22:13.298 [Pipeline] } 00:22:13.314 [Pipeline] // stage 00:22:13.325 [Pipeline] dir 00:22:13.325 Running in /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt 00:22:13.327 [Pipeline] { 00:22:13.341 [Pipeline] catchError 00:22:13.343 [Pipeline] { 00:22:13.358 [Pipeline] sh 00:22:13.635 + vagrant ssh-config --host vagrant 00:22:13.635 + sed -ne '/^Host/,$p' 00:22:13.635 + tee ssh_conf 00:22:16.162 Host vagrant 00:22:16.162 HostName 192.168.121.104 00:22:16.162 User vagrant 00:22:16.162 Port 22 00:22:16.162 UserKnownHostsFile /dev/null 00:22:16.162 StrictHostKeyChecking no 00:22:16.162 PasswordAuthentication no 00:22:16.162 IdentityFile /var/lib/libvirt/images/.vagrant.d/boxes/spdk-VAGRANTSLASH-fedora39/39-1.5-1721788873-2326/libvirt/fedora39 00:22:16.162 IdentitiesOnly yes 00:22:16.162 LogLevel FATAL 00:22:16.162 ForwardAgent yes 00:22:16.162 ForwardX11 yes 00:22:16.162 00:22:16.178 [Pipeline] withEnv 00:22:16.181 [Pipeline] { 00:22:16.196 [Pipeline] sh 00:22:16.556 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant '#!/bin/bash 00:22:16.556 source /etc/os-release 00:22:16.556 [[ -e /image.version ]] && img=$(< /image.version) 00:22:16.556 # Minimal, systemd-like check. 00:22:16.556 if [[ -e /.dockerenv ]]; then 00:22:16.556 # Clear garbage from the node'\''s name: 00:22:16.556 # agt-er_autotest_547-896 -> autotest_547-896 00:22:16.556 # $HOSTNAME is the actual container id 00:22:16.556 agent=$HOSTNAME@${DOCKER_SWARM_PLUGIN_JENKINS_AGENT_NAME#*_} 00:22:16.556 if grep -q "/etc/hostname" /proc/self/mountinfo; then 00:22:16.556 # We can assume this is a mount from a host where container is running, 00:22:16.556 # so fetch its hostname to easily identify the target swarm worker. 00:22:16.556 container="$(< /etc/hostname) ($agent)" 00:22:16.556 else 00:22:16.556 # Fallback 00:22:16.556 container=$agent 00:22:16.556 fi 00:22:16.556 fi 00:22:16.556 echo "${NAME} ${VERSION_ID}|$(uname -r)|${img:-N/A}|${container:-N/A}" 00:22:16.556 ' 00:22:16.566 [Pipeline] } 00:22:16.583 [Pipeline] // withEnv 00:22:16.592 [Pipeline] setCustomBuildProperty 00:22:16.611 [Pipeline] stage 00:22:16.614 [Pipeline] { (Tests) 00:22:16.635 [Pipeline] sh 00:22:16.917 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/jbp/jenkins/jjb-config/jobs/scripts/autoruner.sh vagrant@vagrant:./ 00:22:16.932 [Pipeline] sh 00:22:17.209 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/jbp/jenkins/jjb-config/jobs/scripts/pkgdep-autoruner.sh vagrant@vagrant:./ 00:22:17.227 [Pipeline] timeout 00:22:17.227 Timeout set to expire in 1 hr 30 min 00:22:17.229 [Pipeline] { 00:22:17.245 [Pipeline] sh 00:22:17.523 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant 'git -C spdk_repo/spdk reset --hard' 00:22:17.780 HEAD is now at a361eb5e2 nvme_spec: Add SPDK_NVME_IO_FLAGS_PRCHK_MASK 00:22:18.050 [Pipeline] sh 00:22:18.327 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant 'sudo chown vagrant:vagrant spdk_repo' 00:22:18.339 [Pipeline] sh 00:22:18.616 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf vagrant@vagrant:spdk_repo 00:22:18.632 [Pipeline] sh 00:22:18.971 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant 'JOB_BASE_NAME=raid-vg-autotest ./autoruner.sh spdk_repo' 00:22:18.971 ++ readlink -f spdk_repo 00:22:18.971 + DIR_ROOT=/home/vagrant/spdk_repo 00:22:18.971 + [[ -n /home/vagrant/spdk_repo ]] 00:22:18.971 + DIR_SPDK=/home/vagrant/spdk_repo/spdk 00:22:18.971 + DIR_OUTPUT=/home/vagrant/spdk_repo/output 00:22:18.971 + [[ -d /home/vagrant/spdk_repo/spdk ]] 00:22:18.971 + [[ ! -d /home/vagrant/spdk_repo/output ]] 00:22:18.971 + [[ -d /home/vagrant/spdk_repo/output ]] 00:22:18.971 + [[ raid-vg-autotest == pkgdep-* ]] 00:22:18.971 + cd /home/vagrant/spdk_repo 00:22:18.971 + source /etc/os-release 00:22:18.971 ++ NAME='Fedora Linux' 00:22:18.971 ++ VERSION='39 (Cloud Edition)' 00:22:18.971 ++ ID=fedora 00:22:18.971 ++ VERSION_ID=39 00:22:18.971 ++ VERSION_CODENAME= 00:22:18.971 ++ PLATFORM_ID=platform:f39 00:22:18.971 ++ PRETTY_NAME='Fedora Linux 39 (Cloud Edition)' 00:22:18.971 ++ ANSI_COLOR='0;38;2;60;110;180' 00:22:18.971 ++ LOGO=fedora-logo-icon 00:22:18.971 ++ CPE_NAME=cpe:/o:fedoraproject:fedora:39 00:22:18.971 ++ HOME_URL=https://fedoraproject.org/ 00:22:18.971 ++ DOCUMENTATION_URL=https://docs.fedoraproject.org/en-US/fedora/f39/system-administrators-guide/ 00:22:18.971 ++ SUPPORT_URL=https://ask.fedoraproject.org/ 00:22:18.971 ++ BUG_REPORT_URL=https://bugzilla.redhat.com/ 00:22:18.971 ++ REDHAT_BUGZILLA_PRODUCT=Fedora 00:22:18.971 ++ REDHAT_BUGZILLA_PRODUCT_VERSION=39 00:22:18.971 ++ REDHAT_SUPPORT_PRODUCT=Fedora 00:22:18.971 ++ REDHAT_SUPPORT_PRODUCT_VERSION=39 00:22:18.971 ++ SUPPORT_END=2024-11-12 00:22:18.971 ++ VARIANT='Cloud Edition' 00:22:18.971 ++ VARIANT_ID=cloud 00:22:18.971 + uname -a 00:22:18.971 Linux fedora39-cloud-1721788873-2326 6.8.9-200.fc39.x86_64 #1 SMP PREEMPT_DYNAMIC Wed Jul 24 03:04:40 UTC 2024 x86_64 GNU/Linux 00:22:18.971 + sudo /home/vagrant/spdk_repo/spdk/scripts/setup.sh status 00:22:19.536 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:22:19.536 Hugepages 00:22:19.536 node hugesize free / total 00:22:19.536 node0 1048576kB 0 / 0 00:22:19.536 node0 2048kB 0 / 0 00:22:19.536 00:22:19.536 Type BDF Vendor Device NUMA Driver Device Block devices 00:22:19.536 virtio 0000:00:03.0 1af4 1001 unknown virtio-pci - vda 00:22:19.536 NVMe 0000:00:10.0 1b36 0010 unknown nvme nvme0 nvme0n1 00:22:19.536 NVMe 0000:00:11.0 1b36 0010 unknown nvme nvme1 nvme1n1 nvme1n2 nvme1n3 00:22:19.536 + rm -f /tmp/spdk-ld-path 00:22:19.536 + source autorun-spdk.conf 00:22:19.536 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:22:19.536 ++ SPDK_RUN_ASAN=1 00:22:19.536 ++ SPDK_RUN_UBSAN=1 00:22:19.536 ++ SPDK_TEST_RAID=1 00:22:19.536 ++ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:22:19.536 ++ RUN_NIGHTLY=0 00:22:19.536 + (( SPDK_TEST_NVME_CMB == 1 || SPDK_TEST_NVME_PMR == 1 )) 00:22:19.536 + [[ -n '' ]] 00:22:19.536 + sudo git config --global --add safe.directory /home/vagrant/spdk_repo/spdk 00:22:19.536 + for M in /var/spdk/build-*-manifest.txt 00:22:19.536 + [[ -f /var/spdk/build-kernel-manifest.txt ]] 00:22:19.794 + cp /var/spdk/build-kernel-manifest.txt /home/vagrant/spdk_repo/output/ 00:22:19.794 + for M in /var/spdk/build-*-manifest.txt 00:22:19.794 + [[ -f /var/spdk/build-pkg-manifest.txt ]] 00:22:19.794 + cp /var/spdk/build-pkg-manifest.txt /home/vagrant/spdk_repo/output/ 00:22:19.794 + for M in /var/spdk/build-*-manifest.txt 00:22:19.794 + [[ -f /var/spdk/build-repo-manifest.txt ]] 00:22:19.794 + cp /var/spdk/build-repo-manifest.txt /home/vagrant/spdk_repo/output/ 00:22:19.794 ++ uname 00:22:19.794 + [[ Linux == \L\i\n\u\x ]] 00:22:19.794 + sudo dmesg -T 00:22:19.794 + sudo dmesg --clear 00:22:19.794 + dmesg_pid=4990 00:22:19.794 + sudo dmesg -Tw 00:22:19.794 + [[ Fedora Linux == FreeBSD ]] 00:22:19.794 + export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:22:19.794 + UNBIND_ENTIRE_IOMMU_GROUP=yes 00:22:19.794 + [[ -e /var/spdk/dependencies/vhost/spdk_test_image.qcow2 ]] 00:22:19.794 + [[ -x /usr/src/fio-static/fio ]] 00:22:19.794 + export FIO_BIN=/usr/src/fio-static/fio 00:22:19.794 + FIO_BIN=/usr/src/fio-static/fio 00:22:19.794 + [[ '' == \/\q\e\m\u\_\v\f\i\o\/* ]] 00:22:19.794 + [[ ! -v VFIO_QEMU_BIN ]] 00:22:19.794 + [[ -e /usr/local/qemu/vfio-user-latest ]] 00:22:19.794 + export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:22:19.794 + VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:22:19.794 + [[ -e /usr/local/qemu/vanilla-latest ]] 00:22:19.794 + export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:22:19.794 + QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:22:19.794 + spdk/autorun.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:22:19.794 13:42:26 -- common/autotest_common.sh@1692 -- $ [[ n == y ]] 00:22:19.794 13:42:26 -- spdk/autorun.sh@20 -- $ source /home/vagrant/spdk_repo/autorun-spdk.conf 00:22:19.794 13:42:26 -- spdk_repo/autorun-spdk.conf@1 -- $ SPDK_RUN_FUNCTIONAL_TEST=1 00:22:19.794 13:42:26 -- spdk_repo/autorun-spdk.conf@2 -- $ SPDK_RUN_ASAN=1 00:22:19.794 13:42:26 -- spdk_repo/autorun-spdk.conf@3 -- $ SPDK_RUN_UBSAN=1 00:22:19.794 13:42:26 -- spdk_repo/autorun-spdk.conf@4 -- $ SPDK_TEST_RAID=1 00:22:19.794 13:42:26 -- spdk_repo/autorun-spdk.conf@5 -- $ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:22:19.794 13:42:26 -- spdk_repo/autorun-spdk.conf@6 -- $ RUN_NIGHTLY=0 00:22:19.794 13:42:26 -- spdk/autorun.sh@22 -- $ trap 'timing_finish || exit 1' EXIT 00:22:19.794 13:42:26 -- spdk/autorun.sh@25 -- $ /home/vagrant/spdk_repo/spdk/autobuild.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:22:19.794 13:42:26 -- common/autotest_common.sh@1692 -- $ [[ n == y ]] 00:22:19.794 13:42:26 -- common/autobuild_common.sh@15 -- $ source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:22:19.794 13:42:26 -- scripts/common.sh@15 -- $ shopt -s extglob 00:22:19.794 13:42:26 -- scripts/common.sh@544 -- $ [[ -e /bin/wpdk_common.sh ]] 00:22:19.794 13:42:26 -- scripts/common.sh@552 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:22:19.794 13:42:26 -- scripts/common.sh@553 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:22:19.794 13:42:26 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:19.794 13:42:26 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:19.794 13:42:26 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:19.794 13:42:26 -- paths/export.sh@5 -- $ export PATH 00:22:19.794 13:42:26 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:19.794 13:42:26 -- common/autobuild_common.sh@492 -- $ out=/home/vagrant/spdk_repo/spdk/../output 00:22:19.794 13:42:26 -- common/autobuild_common.sh@493 -- $ date +%s 00:22:19.794 13:42:26 -- common/autobuild_common.sh@493 -- $ mktemp -dt spdk_1732110146.XXXXXX 00:22:19.794 13:42:26 -- common/autobuild_common.sh@493 -- $ SPDK_WORKSPACE=/tmp/spdk_1732110146.eqCLBJ 00:22:19.794 13:42:26 -- common/autobuild_common.sh@495 -- $ [[ -n '' ]] 00:22:19.794 13:42:26 -- common/autobuild_common.sh@499 -- $ '[' -n '' ']' 00:22:19.794 13:42:26 -- common/autobuild_common.sh@502 -- $ scanbuild_exclude='--exclude /home/vagrant/spdk_repo/spdk/dpdk/' 00:22:19.794 13:42:26 -- common/autobuild_common.sh@506 -- $ scanbuild_exclude+=' --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp' 00:22:19.794 13:42:26 -- common/autobuild_common.sh@508 -- $ scanbuild='scan-build -o /home/vagrant/spdk_repo/spdk/../output/scan-build-tmp --exclude /home/vagrant/spdk_repo/spdk/dpdk/ --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp --status-bugs' 00:22:19.794 13:42:26 -- common/autobuild_common.sh@509 -- $ get_config_params 00:22:19.794 13:42:26 -- common/autotest_common.sh@409 -- $ xtrace_disable 00:22:19.794 13:42:26 -- common/autotest_common.sh@10 -- $ set +x 00:22:19.794 13:42:26 -- common/autobuild_common.sh@509 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f' 00:22:19.794 13:42:26 -- common/autobuild_common.sh@511 -- $ start_monitor_resources 00:22:19.794 13:42:26 -- pm/common@17 -- $ local monitor 00:22:19.794 13:42:26 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:22:19.794 13:42:26 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:22:19.794 13:42:26 -- pm/common@25 -- $ sleep 1 00:22:19.794 13:42:26 -- pm/common@21 -- $ date +%s 00:22:19.794 13:42:26 -- pm/common@21 -- $ date +%s 00:22:19.794 13:42:26 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autobuild.sh.1732110146 00:22:19.794 13:42:26 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autobuild.sh.1732110146 00:22:19.794 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autobuild.sh.1732110146_collect-cpu-load.pm.log 00:22:19.794 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autobuild.sh.1732110146_collect-vmstat.pm.log 00:22:21.167 13:42:27 -- common/autobuild_common.sh@512 -- $ trap stop_monitor_resources EXIT 00:22:21.167 13:42:27 -- spdk/autobuild.sh@11 -- $ SPDK_TEST_AUTOBUILD= 00:22:21.167 13:42:27 -- spdk/autobuild.sh@12 -- $ umask 022 00:22:21.167 13:42:27 -- spdk/autobuild.sh@13 -- $ cd /home/vagrant/spdk_repo/spdk 00:22:21.167 13:42:27 -- spdk/autobuild.sh@16 -- $ date -u 00:22:21.167 Wed Nov 20 01:42:27 PM UTC 2024 00:22:21.167 13:42:27 -- spdk/autobuild.sh@17 -- $ git describe --tags 00:22:21.167 v25.01-pre-248-ga361eb5e2 00:22:21.167 13:42:27 -- spdk/autobuild.sh@19 -- $ '[' 1 -eq 1 ']' 00:22:21.167 13:42:27 -- spdk/autobuild.sh@20 -- $ run_test asan echo 'using asan' 00:22:21.167 13:42:27 -- common/autotest_common.sh@1105 -- $ '[' 3 -le 1 ']' 00:22:21.167 13:42:27 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:22:21.167 13:42:27 -- common/autotest_common.sh@10 -- $ set +x 00:22:21.167 ************************************ 00:22:21.167 START TEST asan 00:22:21.167 ************************************ 00:22:21.167 using asan 00:22:21.167 13:42:27 asan -- common/autotest_common.sh@1129 -- $ echo 'using asan' 00:22:21.167 00:22:21.167 real 0m0.000s 00:22:21.167 user 0m0.000s 00:22:21.167 sys 0m0.000s 00:22:21.167 13:42:27 asan -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:22:21.167 13:42:27 asan -- common/autotest_common.sh@10 -- $ set +x 00:22:21.167 ************************************ 00:22:21.167 END TEST asan 00:22:21.167 ************************************ 00:22:21.167 13:42:27 -- spdk/autobuild.sh@23 -- $ '[' 1 -eq 1 ']' 00:22:21.167 13:42:27 -- spdk/autobuild.sh@24 -- $ run_test ubsan echo 'using ubsan' 00:22:21.167 13:42:27 -- common/autotest_common.sh@1105 -- $ '[' 3 -le 1 ']' 00:22:21.167 13:42:27 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:22:21.167 13:42:27 -- common/autotest_common.sh@10 -- $ set +x 00:22:21.167 ************************************ 00:22:21.167 START TEST ubsan 00:22:21.167 ************************************ 00:22:21.167 using ubsan 00:22:21.167 13:42:27 ubsan -- common/autotest_common.sh@1129 -- $ echo 'using ubsan' 00:22:21.167 00:22:21.167 real 0m0.000s 00:22:21.167 user 0m0.000s 00:22:21.167 sys 0m0.000s 00:22:21.167 13:42:27 ubsan -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:22:21.167 ************************************ 00:22:21.167 END TEST ubsan 00:22:21.167 ************************************ 00:22:21.167 13:42:27 ubsan -- common/autotest_common.sh@10 -- $ set +x 00:22:21.167 13:42:27 -- spdk/autobuild.sh@27 -- $ '[' -n '' ']' 00:22:21.167 13:42:27 -- spdk/autobuild.sh@31 -- $ case "$SPDK_TEST_AUTOBUILD" in 00:22:21.167 13:42:27 -- spdk/autobuild.sh@47 -- $ [[ 0 -eq 1 ]] 00:22:21.167 13:42:27 -- spdk/autobuild.sh@51 -- $ [[ 0 -eq 1 ]] 00:22:21.167 13:42:27 -- spdk/autobuild.sh@55 -- $ [[ -n '' ]] 00:22:21.167 13:42:27 -- spdk/autobuild.sh@57 -- $ [[ 0 -eq 1 ]] 00:22:21.167 13:42:27 -- spdk/autobuild.sh@59 -- $ [[ 0 -eq 1 ]] 00:22:21.167 13:42:27 -- spdk/autobuild.sh@62 -- $ [[ 0 -eq 1 ]] 00:22:21.167 13:42:27 -- spdk/autobuild.sh@67 -- $ /home/vagrant/spdk_repo/spdk/configure --enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f --with-shared 00:22:21.167 Using default SPDK env in /home/vagrant/spdk_repo/spdk/lib/env_dpdk 00:22:21.167 Using default DPDK in /home/vagrant/spdk_repo/spdk/dpdk/build 00:22:21.423 Using 'verbs' RDMA provider 00:22:32.001 Configuring ISA-L (logfile: /home/vagrant/spdk_repo/spdk/.spdk-isal.log)...done. 00:22:41.967 Configuring ISA-L-crypto (logfile: /home/vagrant/spdk_repo/spdk/.spdk-isal-crypto.log)...done. 00:22:42.226 Creating mk/config.mk...done. 00:22:42.226 Creating mk/cc.flags.mk...done. 00:22:42.226 Type 'make' to build. 00:22:42.226 13:42:49 -- spdk/autobuild.sh@70 -- $ run_test make make -j10 00:22:42.226 13:42:49 -- common/autotest_common.sh@1105 -- $ '[' 3 -le 1 ']' 00:22:42.226 13:42:49 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:22:42.226 13:42:49 -- common/autotest_common.sh@10 -- $ set +x 00:22:42.226 ************************************ 00:22:42.226 START TEST make 00:22:42.226 ************************************ 00:22:42.226 13:42:49 make -- common/autotest_common.sh@1129 -- $ make -j10 00:22:42.486 make[1]: Nothing to be done for 'all'. 00:22:54.675 The Meson build system 00:22:54.675 Version: 1.5.0 00:22:54.675 Source dir: /home/vagrant/spdk_repo/spdk/dpdk 00:22:54.675 Build dir: /home/vagrant/spdk_repo/spdk/dpdk/build-tmp 00:22:54.675 Build type: native build 00:22:54.675 Program cat found: YES (/usr/bin/cat) 00:22:54.675 Project name: DPDK 00:22:54.675 Project version: 24.03.0 00:22:54.676 C compiler for the host machine: cc (gcc 13.3.1 "cc (GCC) 13.3.1 20240522 (Red Hat 13.3.1-1)") 00:22:54.676 C linker for the host machine: cc ld.bfd 2.40-14 00:22:54.676 Host machine cpu family: x86_64 00:22:54.676 Host machine cpu: x86_64 00:22:54.676 Message: ## Building in Developer Mode ## 00:22:54.676 Program pkg-config found: YES (/usr/bin/pkg-config) 00:22:54.676 Program check-symbols.sh found: YES (/home/vagrant/spdk_repo/spdk/dpdk/buildtools/check-symbols.sh) 00:22:54.676 Program options-ibverbs-static.sh found: YES (/home/vagrant/spdk_repo/spdk/dpdk/buildtools/options-ibverbs-static.sh) 00:22:54.676 Program python3 found: YES (/usr/bin/python3) 00:22:54.676 Program cat found: YES (/usr/bin/cat) 00:22:54.676 Compiler for C supports arguments -march=native: YES 00:22:54.676 Checking for size of "void *" : 8 00:22:54.676 Checking for size of "void *" : 8 (cached) 00:22:54.676 Compiler for C supports link arguments -Wl,--undefined-version: YES 00:22:54.676 Library m found: YES 00:22:54.676 Library numa found: YES 00:22:54.676 Has header "numaif.h" : YES 00:22:54.676 Library fdt found: NO 00:22:54.676 Library execinfo found: NO 00:22:54.676 Has header "execinfo.h" : YES 00:22:54.676 Found pkg-config: YES (/usr/bin/pkg-config) 1.9.5 00:22:54.676 Run-time dependency libarchive found: NO (tried pkgconfig) 00:22:54.676 Run-time dependency libbsd found: NO (tried pkgconfig) 00:22:54.676 Run-time dependency jansson found: NO (tried pkgconfig) 00:22:54.676 Run-time dependency openssl found: YES 3.1.1 00:22:54.676 Run-time dependency libpcap found: YES 1.10.4 00:22:54.676 Has header "pcap.h" with dependency libpcap: YES 00:22:54.676 Compiler for C supports arguments -Wcast-qual: YES 00:22:54.676 Compiler for C supports arguments -Wdeprecated: YES 00:22:54.676 Compiler for C supports arguments -Wformat: YES 00:22:54.676 Compiler for C supports arguments -Wformat-nonliteral: NO 00:22:54.676 Compiler for C supports arguments -Wformat-security: NO 00:22:54.676 Compiler for C supports arguments -Wmissing-declarations: YES 00:22:54.676 Compiler for C supports arguments -Wmissing-prototypes: YES 00:22:54.676 Compiler for C supports arguments -Wnested-externs: YES 00:22:54.676 Compiler for C supports arguments -Wold-style-definition: YES 00:22:54.676 Compiler for C supports arguments -Wpointer-arith: YES 00:22:54.676 Compiler for C supports arguments -Wsign-compare: YES 00:22:54.676 Compiler for C supports arguments -Wstrict-prototypes: YES 00:22:54.676 Compiler for C supports arguments -Wundef: YES 00:22:54.676 Compiler for C supports arguments -Wwrite-strings: YES 00:22:54.676 Compiler for C supports arguments -Wno-address-of-packed-member: YES 00:22:54.676 Compiler for C supports arguments -Wno-packed-not-aligned: YES 00:22:54.676 Compiler for C supports arguments -Wno-missing-field-initializers: YES 00:22:54.676 Compiler for C supports arguments -Wno-zero-length-bounds: YES 00:22:54.676 Program objdump found: YES (/usr/bin/objdump) 00:22:54.676 Compiler for C supports arguments -mavx512f: YES 00:22:54.676 Checking if "AVX512 checking" compiles: YES 00:22:54.676 Fetching value of define "__SSE4_2__" : 1 00:22:54.676 Fetching value of define "__AES__" : 1 00:22:54.676 Fetching value of define "__AVX__" : 1 00:22:54.676 Fetching value of define "__AVX2__" : 1 00:22:54.676 Fetching value of define "__AVX512BW__" : 1 00:22:54.676 Fetching value of define "__AVX512CD__" : 1 00:22:54.676 Fetching value of define "__AVX512DQ__" : 1 00:22:54.676 Fetching value of define "__AVX512F__" : 1 00:22:54.676 Fetching value of define "__AVX512VL__" : 1 00:22:54.676 Fetching value of define "__PCLMUL__" : 1 00:22:54.676 Fetching value of define "__RDRND__" : 1 00:22:54.676 Fetching value of define "__RDSEED__" : 1 00:22:54.676 Fetching value of define "__VPCLMULQDQ__" : 1 00:22:54.676 Fetching value of define "__znver1__" : (undefined) 00:22:54.676 Fetching value of define "__znver2__" : (undefined) 00:22:54.676 Fetching value of define "__znver3__" : (undefined) 00:22:54.676 Fetching value of define "__znver4__" : (undefined) 00:22:54.676 Library asan found: YES 00:22:54.676 Compiler for C supports arguments -Wno-format-truncation: YES 00:22:54.676 Message: lib/log: Defining dependency "log" 00:22:54.676 Message: lib/kvargs: Defining dependency "kvargs" 00:22:54.676 Message: lib/telemetry: Defining dependency "telemetry" 00:22:54.676 Library rt found: YES 00:22:54.676 Checking for function "getentropy" : NO 00:22:54.676 Message: lib/eal: Defining dependency "eal" 00:22:54.676 Message: lib/ring: Defining dependency "ring" 00:22:54.676 Message: lib/rcu: Defining dependency "rcu" 00:22:54.676 Message: lib/mempool: Defining dependency "mempool" 00:22:54.676 Message: lib/mbuf: Defining dependency "mbuf" 00:22:54.676 Fetching value of define "__PCLMUL__" : 1 (cached) 00:22:54.676 Fetching value of define "__AVX512F__" : 1 (cached) 00:22:54.676 Fetching value of define "__AVX512BW__" : 1 (cached) 00:22:54.676 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:22:54.676 Fetching value of define "__AVX512VL__" : 1 (cached) 00:22:54.676 Fetching value of define "__VPCLMULQDQ__" : 1 (cached) 00:22:54.676 Compiler for C supports arguments -mpclmul: YES 00:22:54.676 Compiler for C supports arguments -maes: YES 00:22:54.676 Compiler for C supports arguments -mavx512f: YES (cached) 00:22:54.676 Compiler for C supports arguments -mavx512bw: YES 00:22:54.676 Compiler for C supports arguments -mavx512dq: YES 00:22:54.676 Compiler for C supports arguments -mavx512vl: YES 00:22:54.676 Compiler for C supports arguments -mvpclmulqdq: YES 00:22:54.676 Compiler for C supports arguments -mavx2: YES 00:22:54.676 Compiler for C supports arguments -mavx: YES 00:22:54.676 Message: lib/net: Defining dependency "net" 00:22:54.676 Message: lib/meter: Defining dependency "meter" 00:22:54.676 Message: lib/ethdev: Defining dependency "ethdev" 00:22:54.676 Message: lib/pci: Defining dependency "pci" 00:22:54.676 Message: lib/cmdline: Defining dependency "cmdline" 00:22:54.676 Message: lib/hash: Defining dependency "hash" 00:22:54.676 Message: lib/timer: Defining dependency "timer" 00:22:54.676 Message: lib/compressdev: Defining dependency "compressdev" 00:22:54.676 Message: lib/cryptodev: Defining dependency "cryptodev" 00:22:54.676 Message: lib/dmadev: Defining dependency "dmadev" 00:22:54.676 Compiler for C supports arguments -Wno-cast-qual: YES 00:22:54.676 Message: lib/power: Defining dependency "power" 00:22:54.676 Message: lib/reorder: Defining dependency "reorder" 00:22:54.676 Message: lib/security: Defining dependency "security" 00:22:54.676 Has header "linux/userfaultfd.h" : YES 00:22:54.676 Has header "linux/vduse.h" : YES 00:22:54.676 Message: lib/vhost: Defining dependency "vhost" 00:22:54.676 Compiler for C supports arguments -Wno-format-truncation: YES (cached) 00:22:54.676 Message: drivers/bus/pci: Defining dependency "bus_pci" 00:22:54.676 Message: drivers/bus/vdev: Defining dependency "bus_vdev" 00:22:54.676 Message: drivers/mempool/ring: Defining dependency "mempool_ring" 00:22:54.676 Message: Disabling raw/* drivers: missing internal dependency "rawdev" 00:22:54.676 Message: Disabling regex/* drivers: missing internal dependency "regexdev" 00:22:54.676 Message: Disabling ml/* drivers: missing internal dependency "mldev" 00:22:54.676 Message: Disabling event/* drivers: missing internal dependency "eventdev" 00:22:54.676 Message: Disabling baseband/* drivers: missing internal dependency "bbdev" 00:22:54.676 Message: Disabling gpu/* drivers: missing internal dependency "gpudev" 00:22:54.676 Program doxygen found: YES (/usr/local/bin/doxygen) 00:22:54.676 Configuring doxy-api-html.conf using configuration 00:22:54.676 Configuring doxy-api-man.conf using configuration 00:22:54.676 Program mandb found: YES (/usr/bin/mandb) 00:22:54.676 Program sphinx-build found: NO 00:22:54.676 Configuring rte_build_config.h using configuration 00:22:54.676 Message: 00:22:54.676 ================= 00:22:54.676 Applications Enabled 00:22:54.676 ================= 00:22:54.676 00:22:54.676 apps: 00:22:54.676 00:22:54.676 00:22:54.676 Message: 00:22:54.676 ================= 00:22:54.676 Libraries Enabled 00:22:54.676 ================= 00:22:54.676 00:22:54.676 libs: 00:22:54.676 log, kvargs, telemetry, eal, ring, rcu, mempool, mbuf, 00:22:54.676 net, meter, ethdev, pci, cmdline, hash, timer, compressdev, 00:22:54.676 cryptodev, dmadev, power, reorder, security, vhost, 00:22:54.676 00:22:54.677 Message: 00:22:54.677 =============== 00:22:54.677 Drivers Enabled 00:22:54.677 =============== 00:22:54.677 00:22:54.677 common: 00:22:54.677 00:22:54.677 bus: 00:22:54.677 pci, vdev, 00:22:54.677 mempool: 00:22:54.677 ring, 00:22:54.677 dma: 00:22:54.677 00:22:54.677 net: 00:22:54.677 00:22:54.677 crypto: 00:22:54.677 00:22:54.677 compress: 00:22:54.677 00:22:54.677 vdpa: 00:22:54.677 00:22:54.677 00:22:54.677 Message: 00:22:54.677 ================= 00:22:54.677 Content Skipped 00:22:54.677 ================= 00:22:54.677 00:22:54.677 apps: 00:22:54.677 dumpcap: explicitly disabled via build config 00:22:54.677 graph: explicitly disabled via build config 00:22:54.677 pdump: explicitly disabled via build config 00:22:54.677 proc-info: explicitly disabled via build config 00:22:54.677 test-acl: explicitly disabled via build config 00:22:54.677 test-bbdev: explicitly disabled via build config 00:22:54.677 test-cmdline: explicitly disabled via build config 00:22:54.677 test-compress-perf: explicitly disabled via build config 00:22:54.677 test-crypto-perf: explicitly disabled via build config 00:22:54.677 test-dma-perf: explicitly disabled via build config 00:22:54.677 test-eventdev: explicitly disabled via build config 00:22:54.677 test-fib: explicitly disabled via build config 00:22:54.677 test-flow-perf: explicitly disabled via build config 00:22:54.677 test-gpudev: explicitly disabled via build config 00:22:54.677 test-mldev: explicitly disabled via build config 00:22:54.677 test-pipeline: explicitly disabled via build config 00:22:54.677 test-pmd: explicitly disabled via build config 00:22:54.677 test-regex: explicitly disabled via build config 00:22:54.677 test-sad: explicitly disabled via build config 00:22:54.677 test-security-perf: explicitly disabled via build config 00:22:54.677 00:22:54.677 libs: 00:22:54.677 argparse: explicitly disabled via build config 00:22:54.677 metrics: explicitly disabled via build config 00:22:54.677 acl: explicitly disabled via build config 00:22:54.677 bbdev: explicitly disabled via build config 00:22:54.677 bitratestats: explicitly disabled via build config 00:22:54.677 bpf: explicitly disabled via build config 00:22:54.677 cfgfile: explicitly disabled via build config 00:22:54.677 distributor: explicitly disabled via build config 00:22:54.677 efd: explicitly disabled via build config 00:22:54.677 eventdev: explicitly disabled via build config 00:22:54.677 dispatcher: explicitly disabled via build config 00:22:54.677 gpudev: explicitly disabled via build config 00:22:54.677 gro: explicitly disabled via build config 00:22:54.677 gso: explicitly disabled via build config 00:22:54.677 ip_frag: explicitly disabled via build config 00:22:54.677 jobstats: explicitly disabled via build config 00:22:54.677 latencystats: explicitly disabled via build config 00:22:54.677 lpm: explicitly disabled via build config 00:22:54.677 member: explicitly disabled via build config 00:22:54.677 pcapng: explicitly disabled via build config 00:22:54.677 rawdev: explicitly disabled via build config 00:22:54.677 regexdev: explicitly disabled via build config 00:22:54.677 mldev: explicitly disabled via build config 00:22:54.677 rib: explicitly disabled via build config 00:22:54.677 sched: explicitly disabled via build config 00:22:54.677 stack: explicitly disabled via build config 00:22:54.677 ipsec: explicitly disabled via build config 00:22:54.677 pdcp: explicitly disabled via build config 00:22:54.677 fib: explicitly disabled via build config 00:22:54.677 port: explicitly disabled via build config 00:22:54.677 pdump: explicitly disabled via build config 00:22:54.677 table: explicitly disabled via build config 00:22:54.677 pipeline: explicitly disabled via build config 00:22:54.677 graph: explicitly disabled via build config 00:22:54.677 node: explicitly disabled via build config 00:22:54.677 00:22:54.677 drivers: 00:22:54.677 common/cpt: not in enabled drivers build config 00:22:54.677 common/dpaax: not in enabled drivers build config 00:22:54.677 common/iavf: not in enabled drivers build config 00:22:54.677 common/idpf: not in enabled drivers build config 00:22:54.677 common/ionic: not in enabled drivers build config 00:22:54.677 common/mvep: not in enabled drivers build config 00:22:54.677 common/octeontx: not in enabled drivers build config 00:22:54.677 bus/auxiliary: not in enabled drivers build config 00:22:54.677 bus/cdx: not in enabled drivers build config 00:22:54.677 bus/dpaa: not in enabled drivers build config 00:22:54.677 bus/fslmc: not in enabled drivers build config 00:22:54.677 bus/ifpga: not in enabled drivers build config 00:22:54.677 bus/platform: not in enabled drivers build config 00:22:54.677 bus/uacce: not in enabled drivers build config 00:22:54.677 bus/vmbus: not in enabled drivers build config 00:22:54.677 common/cnxk: not in enabled drivers build config 00:22:54.677 common/mlx5: not in enabled drivers build config 00:22:54.677 common/nfp: not in enabled drivers build config 00:22:54.677 common/nitrox: not in enabled drivers build config 00:22:54.677 common/qat: not in enabled drivers build config 00:22:54.677 common/sfc_efx: not in enabled drivers build config 00:22:54.677 mempool/bucket: not in enabled drivers build config 00:22:54.677 mempool/cnxk: not in enabled drivers build config 00:22:54.677 mempool/dpaa: not in enabled drivers build config 00:22:54.677 mempool/dpaa2: not in enabled drivers build config 00:22:54.677 mempool/octeontx: not in enabled drivers build config 00:22:54.677 mempool/stack: not in enabled drivers build config 00:22:54.677 dma/cnxk: not in enabled drivers build config 00:22:54.677 dma/dpaa: not in enabled drivers build config 00:22:54.677 dma/dpaa2: not in enabled drivers build config 00:22:54.677 dma/hisilicon: not in enabled drivers build config 00:22:54.677 dma/idxd: not in enabled drivers build config 00:22:54.677 dma/ioat: not in enabled drivers build config 00:22:54.677 dma/skeleton: not in enabled drivers build config 00:22:54.677 net/af_packet: not in enabled drivers build config 00:22:54.677 net/af_xdp: not in enabled drivers build config 00:22:54.677 net/ark: not in enabled drivers build config 00:22:54.677 net/atlantic: not in enabled drivers build config 00:22:54.677 net/avp: not in enabled drivers build config 00:22:54.677 net/axgbe: not in enabled drivers build config 00:22:54.677 net/bnx2x: not in enabled drivers build config 00:22:54.677 net/bnxt: not in enabled drivers build config 00:22:54.677 net/bonding: not in enabled drivers build config 00:22:54.677 net/cnxk: not in enabled drivers build config 00:22:54.677 net/cpfl: not in enabled drivers build config 00:22:54.677 net/cxgbe: not in enabled drivers build config 00:22:54.677 net/dpaa: not in enabled drivers build config 00:22:54.677 net/dpaa2: not in enabled drivers build config 00:22:54.677 net/e1000: not in enabled drivers build config 00:22:54.677 net/ena: not in enabled drivers build config 00:22:54.677 net/enetc: not in enabled drivers build config 00:22:54.677 net/enetfec: not in enabled drivers build config 00:22:54.677 net/enic: not in enabled drivers build config 00:22:54.677 net/failsafe: not in enabled drivers build config 00:22:54.677 net/fm10k: not in enabled drivers build config 00:22:54.677 net/gve: not in enabled drivers build config 00:22:54.677 net/hinic: not in enabled drivers build config 00:22:54.678 net/hns3: not in enabled drivers build config 00:22:54.678 net/i40e: not in enabled drivers build config 00:22:54.678 net/iavf: not in enabled drivers build config 00:22:54.678 net/ice: not in enabled drivers build config 00:22:54.678 net/idpf: not in enabled drivers build config 00:22:54.678 net/igc: not in enabled drivers build config 00:22:54.678 net/ionic: not in enabled drivers build config 00:22:54.678 net/ipn3ke: not in enabled drivers build config 00:22:54.678 net/ixgbe: not in enabled drivers build config 00:22:54.678 net/mana: not in enabled drivers build config 00:22:54.678 net/memif: not in enabled drivers build config 00:22:54.678 net/mlx4: not in enabled drivers build config 00:22:54.678 net/mlx5: not in enabled drivers build config 00:22:54.678 net/mvneta: not in enabled drivers build config 00:22:54.678 net/mvpp2: not in enabled drivers build config 00:22:54.678 net/netvsc: not in enabled drivers build config 00:22:54.678 net/nfb: not in enabled drivers build config 00:22:54.678 net/nfp: not in enabled drivers build config 00:22:54.678 net/ngbe: not in enabled drivers build config 00:22:54.678 net/null: not in enabled drivers build config 00:22:54.678 net/octeontx: not in enabled drivers build config 00:22:54.678 net/octeon_ep: not in enabled drivers build config 00:22:54.678 net/pcap: not in enabled drivers build config 00:22:54.678 net/pfe: not in enabled drivers build config 00:22:54.678 net/qede: not in enabled drivers build config 00:22:54.678 net/ring: not in enabled drivers build config 00:22:54.678 net/sfc: not in enabled drivers build config 00:22:54.678 net/softnic: not in enabled drivers build config 00:22:54.678 net/tap: not in enabled drivers build config 00:22:54.678 net/thunderx: not in enabled drivers build config 00:22:54.678 net/txgbe: not in enabled drivers build config 00:22:54.678 net/vdev_netvsc: not in enabled drivers build config 00:22:54.678 net/vhost: not in enabled drivers build config 00:22:54.678 net/virtio: not in enabled drivers build config 00:22:54.678 net/vmxnet3: not in enabled drivers build config 00:22:54.678 raw/*: missing internal dependency, "rawdev" 00:22:54.678 crypto/armv8: not in enabled drivers build config 00:22:54.678 crypto/bcmfs: not in enabled drivers build config 00:22:54.678 crypto/caam_jr: not in enabled drivers build config 00:22:54.678 crypto/ccp: not in enabled drivers build config 00:22:54.678 crypto/cnxk: not in enabled drivers build config 00:22:54.678 crypto/dpaa_sec: not in enabled drivers build config 00:22:54.678 crypto/dpaa2_sec: not in enabled drivers build config 00:22:54.678 crypto/ipsec_mb: not in enabled drivers build config 00:22:54.678 crypto/mlx5: not in enabled drivers build config 00:22:54.678 crypto/mvsam: not in enabled drivers build config 00:22:54.678 crypto/nitrox: not in enabled drivers build config 00:22:54.678 crypto/null: not in enabled drivers build config 00:22:54.678 crypto/octeontx: not in enabled drivers build config 00:22:54.678 crypto/openssl: not in enabled drivers build config 00:22:54.678 crypto/scheduler: not in enabled drivers build config 00:22:54.678 crypto/uadk: not in enabled drivers build config 00:22:54.678 crypto/virtio: not in enabled drivers build config 00:22:54.678 compress/isal: not in enabled drivers build config 00:22:54.678 compress/mlx5: not in enabled drivers build config 00:22:54.678 compress/nitrox: not in enabled drivers build config 00:22:54.678 compress/octeontx: not in enabled drivers build config 00:22:54.678 compress/zlib: not in enabled drivers build config 00:22:54.678 regex/*: missing internal dependency, "regexdev" 00:22:54.678 ml/*: missing internal dependency, "mldev" 00:22:54.678 vdpa/ifc: not in enabled drivers build config 00:22:54.678 vdpa/mlx5: not in enabled drivers build config 00:22:54.678 vdpa/nfp: not in enabled drivers build config 00:22:54.678 vdpa/sfc: not in enabled drivers build config 00:22:54.678 event/*: missing internal dependency, "eventdev" 00:22:54.678 baseband/*: missing internal dependency, "bbdev" 00:22:54.678 gpu/*: missing internal dependency, "gpudev" 00:22:54.678 00:22:54.678 00:22:54.678 Build targets in project: 84 00:22:54.678 00:22:54.678 DPDK 24.03.0 00:22:54.678 00:22:54.678 User defined options 00:22:54.678 buildtype : debug 00:22:54.678 default_library : shared 00:22:54.678 libdir : lib 00:22:54.678 prefix : /home/vagrant/spdk_repo/spdk/dpdk/build 00:22:54.678 b_sanitize : address 00:22:54.678 c_args : -Wno-stringop-overflow -fcommon -Wno-stringop-overread -Wno-array-bounds -fPIC -Werror 00:22:54.678 c_link_args : 00:22:54.678 cpu_instruction_set: native 00:22:54.678 disable_apps : dumpcap,graph,pdump,proc-info,test-acl,test-bbdev,test-cmdline,test-compress-perf,test-crypto-perf,test-dma-perf,test-eventdev,test-fib,test-flow-perf,test-gpudev,test-mldev,test-pipeline,test-pmd,test-regex,test-sad,test-security-perf,test 00:22:54.678 disable_libs : acl,argparse,bbdev,bitratestats,bpf,cfgfile,dispatcher,distributor,efd,eventdev,fib,gpudev,graph,gro,gso,ip_frag,ipsec,jobstats,latencystats,lpm,member,metrics,mldev,node,pcapng,pdcp,pdump,pipeline,port,rawdev,regexdev,rib,sched,stack,table 00:22:54.678 enable_docs : false 00:22:54.678 enable_drivers : bus,bus/pci,bus/vdev,mempool/ring,power/acpi,power/amd_pstate,power/cppc,power/intel_pstate,power/intel_uncore,power/kvm_vm 00:22:54.678 enable_kmods : false 00:22:54.678 max_lcores : 128 00:22:54.678 tests : false 00:22:54.678 00:22:54.678 Found ninja-1.11.1.git.kitware.jobserver-1 at /usr/local/bin/ninja 00:22:54.678 ninja: Entering directory `/home/vagrant/spdk_repo/spdk/dpdk/build-tmp' 00:22:54.678 [1/267] Compiling C object lib/librte_log.a.p/log_log.c.o 00:22:54.678 [2/267] Compiling C object lib/librte_log.a.p/log_log_linux.c.o 00:22:54.678 [3/267] Linking static target lib/librte_log.a 00:22:54.678 [4/267] Compiling C object lib/librte_kvargs.a.p/kvargs_rte_kvargs.c.o 00:22:54.678 [5/267] Linking static target lib/librte_kvargs.a 00:22:54.678 [6/267] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_data.c.o 00:22:54.936 [7/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_class.c.o 00:22:54.936 [8/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_debug.c.o 00:22:54.936 [9/267] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_legacy.c.o 00:22:54.936 [10/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_errno.c.o 00:22:54.936 [11/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hexdump.c.o 00:22:55.194 [12/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_config.c.o 00:22:55.194 [13/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_bus.c.o 00:22:55.194 [14/267] Generating lib/kvargs.sym_chk with a custom command (wrapped by meson to capture output) 00:22:55.194 [15/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_devargs.c.o 00:22:55.194 [16/267] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry.c.o 00:22:55.194 [17/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dev.c.o 00:22:55.194 [18/267] Linking static target lib/librte_telemetry.a 00:22:55.452 [19/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_interrupts.c.o 00:22:55.452 [20/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_mcfg.c.o 00:22:55.452 [21/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_fbarray.c.o 00:22:55.452 [22/267] Generating lib/log.sym_chk with a custom command (wrapped by meson to capture output) 00:22:55.452 [23/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_string_fns.c.o 00:22:55.710 [24/267] Linking target lib/librte_log.so.24.1 00:22:55.710 [25/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_launch.c.o 00:22:55.710 [26/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memalloc.c.o 00:22:55.710 [27/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_lcore.c.o 00:22:55.710 [28/267] Generating symbol file lib/librte_log.so.24.1.p/librte_log.so.24.1.symbols 00:22:55.968 [29/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_uuid.c.o 00:22:55.968 [30/267] Linking target lib/librte_kvargs.so.24.1 00:22:55.968 [31/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memzone.c.o 00:22:55.968 [32/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_tailqs.c.o 00:22:55.968 [33/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memory.c.o 00:22:55.968 [34/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_points.c.o 00:22:55.968 [35/267] Generating symbol file lib/librte_kvargs.so.24.1.p/librte_kvargs.so.24.1.symbols 00:22:55.968 [36/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_reciprocal.c.o 00:22:55.968 [37/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_timer.c.o 00:22:56.224 [38/267] Generating lib/telemetry.sym_chk with a custom command (wrapped by meson to capture output) 00:22:56.224 [39/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_version.c.o 00:22:56.224 [40/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_thread.c.o 00:22:56.224 [41/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_cpuflags.c.o 00:22:56.224 [42/267] Linking target lib/librte_telemetry.so.24.1 00:22:56.224 [43/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hypervisor.c.o 00:22:56.224 [44/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_options.c.o 00:22:56.482 [45/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dynmem.c.o 00:22:56.482 [46/267] Compiling C object lib/librte_eal.a.p/eal_common_malloc_elem.c.o 00:22:56.482 [47/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_random.c.o 00:22:56.482 [48/267] Generating symbol file lib/librte_telemetry.so.24.1.p/librte_telemetry.so.24.1.symbols 00:22:56.482 [49/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_service.c.o 00:22:56.739 [50/267] Compiling C object lib/librte_eal.a.p/eal_common_malloc_heap.c.o 00:22:56.739 [51/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_debug.c.o 00:22:56.739 [52/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace.c.o 00:22:56.739 [53/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_malloc.c.o 00:22:56.739 [54/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_proc.c.o 00:22:56.739 [55/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_ctf.c.o 00:22:56.997 [56/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_utils.c.o 00:22:56.997 [57/267] Compiling C object lib/librte_eal.a.p/eal_common_malloc_mp.c.o 00:22:56.997 [58/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_keepalive.c.o 00:22:56.997 [59/267] Compiling C object lib/librte_eal.a.p/eal_common_hotplug_mp.c.o 00:22:56.997 [60/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_file.c.o 00:22:56.997 [61/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_filesystem.c.o 00:22:56.997 [62/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_cpuflags.c.o 00:22:56.997 [63/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_thread.c.o 00:22:56.997 [64/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_timer.c.o 00:22:56.997 [65/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_firmware.c.o 00:22:57.255 [66/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_memory.c.o 00:22:57.255 [67/267] Compiling C object lib/librte_eal.a.p/eal_unix_rte_thread.c.o 00:22:57.255 [68/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_alarm.c.o 00:22:57.255 [69/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_lcore.c.o 00:22:57.255 [70/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_hugepage_info.c.o 00:22:57.512 [71/267] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cpuflags.c.o 00:22:57.512 [72/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_thread.c.o 00:22:57.512 [73/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_dev.c.o 00:22:57.512 [74/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal.c.o 00:22:57.512 [75/267] Compiling C object lib/librte_eal.a.p/eal_x86_rte_hypervisor.c.o 00:22:57.512 [76/267] Compiling C object lib/librte_eal.a.p/eal_x86_rte_spinlock.c.o 00:22:57.512 [77/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio_mp_sync.c.o 00:22:57.769 [78/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio.c.o 00:22:57.769 [79/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_timer.c.o 00:22:57.769 [80/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memalloc.c.o 00:22:57.769 [81/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memory.c.o 00:22:57.769 [82/267] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cycles.c.o 00:22:57.769 [83/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_interrupts.c.o 00:22:57.769 [84/267] Compiling C object lib/librte_ring.a.p/ring_rte_ring.c.o 00:22:57.769 [85/267] Linking static target lib/librte_ring.a 00:22:58.027 [86/267] Compiling C object lib/librte_eal.a.p/eal_x86_rte_power_intrinsics.c.o 00:22:58.027 [87/267] Compiling C object lib/librte_mempool.a.p/mempool_mempool_trace_points.c.o 00:22:58.027 [88/267] Linking static target lib/librte_eal.a 00:22:58.027 [89/267] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_pool_ops.c.o 00:22:58.027 [90/267] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool.c.o 00:22:58.027 [91/267] Compiling C object lib/librte_rcu.a.p/rcu_rte_rcu_qsbr.c.o 00:22:58.027 [92/267] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops.c.o 00:22:58.027 [93/267] Linking static target lib/librte_rcu.a 00:22:58.027 [94/267] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_ptype.c.o 00:22:58.284 [95/267] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops_default.c.o 00:22:58.284 [96/267] Linking static target lib/librte_mempool.a 00:22:58.284 [97/267] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_dyn.c.o 00:22:58.284 [98/267] Generating lib/ring.sym_chk with a custom command (wrapped by meson to capture output) 00:22:58.284 [99/267] Compiling C object lib/librte_net.a.p/net_rte_net_crc.c.o 00:22:58.540 [100/267] Compiling C object lib/librte_net.a.p/net_rte_ether.c.o 00:22:58.541 [101/267] Compiling C object lib/librte_net.a.p/net_rte_net.c.o 00:22:58.541 [102/267] Compiling C object lib/librte_net.a.p/net_rte_arp.c.o 00:22:58.541 [103/267] Generating lib/rcu.sym_chk with a custom command (wrapped by meson to capture output) 00:22:58.541 [104/267] Compiling C object lib/librte_meter.a.p/meter_rte_meter.c.o 00:22:58.859 [105/267] Linking static target lib/librte_meter.a 00:22:58.859 [106/267] Compiling C object lib/librte_net.a.p/net_net_crc_sse.c.o 00:22:58.859 [107/267] Compiling C object lib/librte_net.a.p/net_net_crc_avx512.c.o 00:22:58.859 [108/267] Linking static target lib/librte_net.a 00:22:58.859 [109/267] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_profile.c.o 00:22:58.859 [110/267] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_private.c.o 00:22:58.859 [111/267] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf.c.o 00:22:58.859 [112/267] Linking static target lib/librte_mbuf.a 00:22:58.859 [113/267] Generating lib/meter.sym_chk with a custom command (wrapped by meson to capture output) 00:22:59.153 [114/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_class_eth.c.o 00:22:59.153 [115/267] Generating lib/net.sym_chk with a custom command (wrapped by meson to capture output) 00:22:59.412 [116/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_cman.c.o 00:22:59.412 [117/267] Generating lib/mempool.sym_chk with a custom command (wrapped by meson to capture output) 00:22:59.412 [118/267] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_driver.c.o 00:22:59.412 [119/267] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_trace_points.c.o 00:22:59.412 [120/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_telemetry.c.o 00:22:59.670 [121/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_mtr.c.o 00:22:59.670 [122/267] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8079.c.o 00:22:59.670 [123/267] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_telemetry.c.o 00:22:59.928 [124/267] Generating lib/mbuf.sym_chk with a custom command (wrapped by meson to capture output) 00:22:59.928 [125/267] Compiling C object lib/librte_pci.a.p/pci_rte_pci.c.o 00:22:59.928 [126/267] Linking static target lib/librte_pci.a 00:22:59.928 [127/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline.c.o 00:22:59.928 [128/267] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8472.c.o 00:22:59.928 [129/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse.c.o 00:22:59.928 [130/267] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_linux_ethtool.c.o 00:23:00.186 [131/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_cirbuf.c.o 00:23:00.186 [132/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_tm.c.o 00:23:00.186 [133/267] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_common.c.o 00:23:00.186 [134/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_num.c.o 00:23:00.186 [135/267] Generating lib/pci.sym_chk with a custom command (wrapped by meson to capture output) 00:23:00.186 [136/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_string.c.o 00:23:00.186 [137/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_portlist.c.o 00:23:00.186 [138/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_socket.c.o 00:23:00.186 [139/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_vt100.c.o 00:23:00.186 [140/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_os_unix.c.o 00:23:00.444 [141/267] Compiling C object lib/librte_hash.a.p/hash_rte_hash_crc.c.o 00:23:00.444 [142/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_ipaddr.c.o 00:23:00.444 [143/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_rdline.c.o 00:23:00.444 [144/267] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8636.c.o 00:23:00.444 [145/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_etheraddr.c.o 00:23:00.444 [146/267] Linking static target lib/librte_cmdline.a 00:23:00.702 [147/267] Compiling C object lib/librte_timer.a.p/timer_rte_timer.c.o 00:23:00.702 [148/267] Linking static target lib/librte_timer.a 00:23:00.702 [149/267] Compiling C object lib/librte_hash.a.p/hash_rte_thash_gfni.c.o 00:23:00.702 [150/267] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev.c.o 00:23:00.961 [151/267] Compiling C object lib/librte_hash.a.p/hash_rte_fbk_hash.c.o 00:23:00.961 [152/267] Compiling C object lib/librte_hash.a.p/hash_rte_thash.c.o 00:23:01.219 [153/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_flow.c.o 00:23:01.219 [154/267] Generating lib/timer.sym_chk with a custom command (wrapped by meson to capture output) 00:23:01.219 [155/267] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev_pmd.c.o 00:23:01.219 [156/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev.c.o 00:23:01.219 [157/267] Linking static target lib/librte_ethdev.a 00:23:01.219 [158/267] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_comp.c.o 00:23:01.219 [159/267] Compiling C object lib/librte_hash.a.p/hash_rte_cuckoo_hash.c.o 00:23:01.219 [160/267] Linking static target lib/librte_compressdev.a 00:23:01.477 [161/267] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_pmd.c.o 00:23:01.477 [162/267] Linking static target lib/librte_hash.a 00:23:01.477 [163/267] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_trace_points.c.o 00:23:01.477 [164/267] Compiling C object lib/librte_power.a.p/power_guest_channel.c.o 00:23:01.735 [165/267] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev_trace_points.c.o 00:23:01.735 [166/267] Compiling C object lib/librte_power.a.p/power_power_acpi_cpufreq.c.o 00:23:01.735 [167/267] Compiling C object lib/librte_power.a.p/power_power_common.c.o 00:23:01.735 [168/267] Compiling C object lib/librte_power.a.p/power_power_kvm_vm.c.o 00:23:01.735 [169/267] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev.c.o 00:23:01.735 [170/267] Linking static target lib/librte_dmadev.a 00:23:01.735 [171/267] Generating lib/cmdline.sym_chk with a custom command (wrapped by meson to capture output) 00:23:01.735 [172/267] Compiling C object lib/librte_power.a.p/power_power_amd_pstate_cpufreq.c.o 00:23:01.993 [173/267] Compiling C object lib/librte_power.a.p/power_power_intel_uncore.c.o 00:23:01.993 [174/267] Compiling C object lib/librte_power.a.p/power_rte_power.c.o 00:23:02.278 [175/267] Compiling C object lib/librte_power.a.p/power_power_cppc_cpufreq.c.o 00:23:02.278 [176/267] Generating lib/compressdev.sym_chk with a custom command (wrapped by meson to capture output) 00:23:02.278 [177/267] Compiling C object lib/librte_power.a.p/power_power_pstate_cpufreq.c.o 00:23:02.278 [178/267] Compiling C object lib/librte_power.a.p/power_rte_power_uncore.c.o 00:23:02.278 [179/267] Generating lib/hash.sym_chk with a custom command (wrapped by meson to capture output) 00:23:02.536 [180/267] Compiling C object lib/librte_vhost.a.p/vhost_fd_man.c.o 00:23:02.536 [181/267] Generating lib/dmadev.sym_chk with a custom command (wrapped by meson to capture output) 00:23:02.536 [182/267] Compiling C object lib/librte_power.a.p/power_rte_power_pmd_mgmt.c.o 00:23:02.536 [183/267] Linking static target lib/librte_power.a 00:23:02.793 [184/267] Compiling C object lib/librte_security.a.p/security_rte_security.c.o 00:23:02.793 [185/267] Linking static target lib/librte_security.a 00:23:02.793 [186/267] Compiling C object lib/librte_cryptodev.a.p/cryptodev_rte_cryptodev.c.o 00:23:02.793 [187/267] Linking static target lib/librte_cryptodev.a 00:23:02.793 [188/267] Compiling C object lib/librte_vhost.a.p/vhost_socket.c.o 00:23:02.793 [189/267] Compiling C object lib/librte_vhost.a.p/vhost_iotlb.c.o 00:23:02.793 [190/267] Compiling C object lib/librte_reorder.a.p/reorder_rte_reorder.c.o 00:23:02.793 [191/267] Linking static target lib/librte_reorder.a 00:23:03.051 [192/267] Compiling C object lib/librte_vhost.a.p/vhost_vdpa.c.o 00:23:03.308 [193/267] Generating lib/security.sym_chk with a custom command (wrapped by meson to capture output) 00:23:03.308 [194/267] Generating lib/reorder.sym_chk with a custom command (wrapped by meson to capture output) 00:23:03.565 [195/267] Compiling C object lib/librte_vhost.a.p/vhost_vhost.c.o 00:23:03.823 [196/267] Generating lib/power.sym_chk with a custom command (wrapped by meson to capture output) 00:23:03.823 [197/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_params.c.o 00:23:03.823 [198/267] Compiling C object lib/librte_vhost.a.p/vhost_vhost_user.c.o 00:23:03.823 [199/267] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net_ctrl.c.o 00:23:04.080 [200/267] Compiling C object lib/librte_vhost.a.p/vhost_vduse.c.o 00:23:04.080 [201/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common.c.o 00:23:04.080 [202/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common_uio.c.o 00:23:04.337 [203/267] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev_params.c.o 00:23:04.337 [204/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci.c.o 00:23:04.337 [205/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_uio.c.o 00:23:04.593 [206/267] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev.c.o 00:23:04.593 [207/267] Linking static target drivers/libtmp_rte_bus_vdev.a 00:23:04.593 [208/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_vfio.c.o 00:23:04.593 [209/267] Linking static target drivers/libtmp_rte_bus_pci.a 00:23:04.851 [210/267] Compiling C object drivers/libtmp_rte_mempool_ring.a.p/mempool_ring_rte_mempool_ring.c.o 00:23:04.851 [211/267] Generating drivers/rte_bus_vdev.pmd.c with a custom command 00:23:04.851 [212/267] Linking static target drivers/libtmp_rte_mempool_ring.a 00:23:04.851 [213/267] Compiling C object drivers/librte_bus_vdev.so.24.1.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:23:04.851 [214/267] Compiling C object drivers/librte_bus_vdev.a.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:23:04.851 [215/267] Linking static target drivers/librte_bus_vdev.a 00:23:04.851 [216/267] Generating drivers/rte_bus_pci.pmd.c with a custom command 00:23:04.851 [217/267] Compiling C object drivers/librte_bus_pci.so.24.1.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:23:04.851 [218/267] Compiling C object drivers/librte_bus_pci.a.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:23:04.851 [219/267] Linking static target drivers/librte_bus_pci.a 00:23:04.851 [220/267] Generating lib/cryptodev.sym_chk with a custom command (wrapped by meson to capture output) 00:23:04.851 [221/267] Generating drivers/rte_mempool_ring.pmd.c with a custom command 00:23:05.108 [222/267] Compiling C object drivers/librte_mempool_ring.so.24.1.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:23:05.108 [223/267] Compiling C object drivers/librte_mempool_ring.a.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:23:05.108 [224/267] Linking static target drivers/librte_mempool_ring.a 00:23:05.108 [225/267] Generating drivers/rte_bus_vdev.sym_chk with a custom command (wrapped by meson to capture output) 00:23:05.365 [226/267] Generating drivers/rte_bus_pci.sym_chk with a custom command (wrapped by meson to capture output) 00:23:05.622 [227/267] Compiling C object lib/librte_vhost.a.p/vhost_vhost_crypto.c.o 00:23:06.555 [228/267] Generating lib/eal.sym_chk with a custom command (wrapped by meson to capture output) 00:23:06.555 [229/267] Linking target lib/librte_eal.so.24.1 00:23:06.813 [230/267] Generating symbol file lib/librte_eal.so.24.1.p/librte_eal.so.24.1.symbols 00:23:06.813 [231/267] Linking target lib/librte_timer.so.24.1 00:23:06.813 [232/267] Linking target lib/librte_ring.so.24.1 00:23:06.813 [233/267] Linking target lib/librte_pci.so.24.1 00:23:06.813 [234/267] Linking target lib/librte_meter.so.24.1 00:23:06.813 [235/267] Linking target lib/librte_dmadev.so.24.1 00:23:06.813 [236/267] Linking target drivers/librte_bus_vdev.so.24.1 00:23:06.813 [237/267] Generating symbol file lib/librte_ring.so.24.1.p/librte_ring.so.24.1.symbols 00:23:06.813 [238/267] Linking target lib/librte_rcu.so.24.1 00:23:06.813 [239/267] Generating symbol file lib/librte_pci.so.24.1.p/librte_pci.so.24.1.symbols 00:23:06.813 [240/267] Generating symbol file lib/librte_timer.so.24.1.p/librte_timer.so.24.1.symbols 00:23:06.813 [241/267] Generating symbol file lib/librte_dmadev.so.24.1.p/librte_dmadev.so.24.1.symbols 00:23:06.813 [242/267] Linking target lib/librte_mempool.so.24.1 00:23:06.813 [243/267] Generating symbol file lib/librte_meter.so.24.1.p/librte_meter.so.24.1.symbols 00:23:06.813 [244/267] Linking target drivers/librte_bus_pci.so.24.1 00:23:07.071 [245/267] Generating symbol file lib/librte_rcu.so.24.1.p/librte_rcu.so.24.1.symbols 00:23:07.071 [246/267] Generating symbol file lib/librte_mempool.so.24.1.p/librte_mempool.so.24.1.symbols 00:23:07.071 [247/267] Linking target drivers/librte_mempool_ring.so.24.1 00:23:07.071 [248/267] Linking target lib/librte_mbuf.so.24.1 00:23:07.071 [249/267] Generating symbol file lib/librte_mbuf.so.24.1.p/librte_mbuf.so.24.1.symbols 00:23:07.071 [250/267] Linking target lib/librte_compressdev.so.24.1 00:23:07.071 [251/267] Linking target lib/librte_reorder.so.24.1 00:23:07.071 [252/267] Linking target lib/librte_net.so.24.1 00:23:07.071 [253/267] Linking target lib/librte_cryptodev.so.24.1 00:23:07.327 [254/267] Generating symbol file lib/librte_net.so.24.1.p/librte_net.so.24.1.symbols 00:23:07.327 [255/267] Generating symbol file lib/librte_cryptodev.so.24.1.p/librte_cryptodev.so.24.1.symbols 00:23:07.327 [256/267] Linking target lib/librte_cmdline.so.24.1 00:23:07.327 [257/267] Linking target lib/librte_hash.so.24.1 00:23:07.328 [258/267] Linking target lib/librte_security.so.24.1 00:23:07.585 [259/267] Generating symbol file lib/librte_hash.so.24.1.p/librte_hash.so.24.1.symbols 00:23:07.585 [260/267] Generating lib/ethdev.sym_chk with a custom command (wrapped by meson to capture output) 00:23:07.585 [261/267] Linking target lib/librte_ethdev.so.24.1 00:23:07.842 [262/267] Generating symbol file lib/librte_ethdev.so.24.1.p/librte_ethdev.so.24.1.symbols 00:23:07.842 [263/267] Linking target lib/librte_power.so.24.1 00:23:08.484 [264/267] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net.c.o 00:23:08.484 [265/267] Linking static target lib/librte_vhost.a 00:23:09.874 [266/267] Generating lib/vhost.sym_chk with a custom command (wrapped by meson to capture output) 00:23:09.874 [267/267] Linking target lib/librte_vhost.so.24.1 00:23:09.874 INFO: autodetecting backend as ninja 00:23:09.874 INFO: calculating backend command to run: /usr/local/bin/ninja -C /home/vagrant/spdk_repo/spdk/dpdk/build-tmp -j 10 00:23:24.742 CC lib/ut_mock/mock.o 00:23:24.742 CC lib/log/log_flags.o 00:23:24.742 CC lib/log/log.o 00:23:24.742 CC lib/log/log_deprecated.o 00:23:24.742 CC lib/ut/ut.o 00:23:24.742 LIB libspdk_ut.a 00:23:24.742 LIB libspdk_ut_mock.a 00:23:24.742 SO libspdk_ut.so.2.0 00:23:24.742 LIB libspdk_log.a 00:23:24.742 SO libspdk_ut_mock.so.6.0 00:23:24.742 SO libspdk_log.so.7.1 00:23:24.742 SYMLINK libspdk_ut.so 00:23:24.742 SYMLINK libspdk_ut_mock.so 00:23:24.742 SYMLINK libspdk_log.so 00:23:24.742 CC lib/ioat/ioat.o 00:23:24.742 CC lib/util/base64.o 00:23:24.742 CC lib/util/cpuset.o 00:23:24.742 CC lib/util/bit_array.o 00:23:24.742 CC lib/dma/dma.o 00:23:24.742 CC lib/util/crc16.o 00:23:24.742 CC lib/util/crc32.o 00:23:24.742 CC lib/util/crc32c.o 00:23:24.742 CXX lib/trace_parser/trace.o 00:23:24.742 CC lib/vfio_user/host/vfio_user_pci.o 00:23:24.742 CC lib/util/crc32_ieee.o 00:23:24.742 CC lib/util/crc64.o 00:23:24.742 CC lib/util/dif.o 00:23:24.742 CC lib/vfio_user/host/vfio_user.o 00:23:24.742 LIB libspdk_dma.a 00:23:24.742 CC lib/util/fd.o 00:23:24.742 SO libspdk_dma.so.5.0 00:23:24.742 CC lib/util/fd_group.o 00:23:24.742 CC lib/util/file.o 00:23:24.742 CC lib/util/hexlify.o 00:23:24.742 SYMLINK libspdk_dma.so 00:23:24.742 CC lib/util/iov.o 00:23:24.742 LIB libspdk_ioat.a 00:23:24.742 SO libspdk_ioat.so.7.0 00:23:24.742 CC lib/util/math.o 00:23:24.742 CC lib/util/net.o 00:23:24.742 SYMLINK libspdk_ioat.so 00:23:24.742 CC lib/util/pipe.o 00:23:24.742 LIB libspdk_vfio_user.a 00:23:24.742 CC lib/util/strerror_tls.o 00:23:24.742 CC lib/util/string.o 00:23:24.742 SO libspdk_vfio_user.so.5.0 00:23:24.742 CC lib/util/uuid.o 00:23:24.742 CC lib/util/xor.o 00:23:24.742 SYMLINK libspdk_vfio_user.so 00:23:24.742 CC lib/util/zipf.o 00:23:24.742 CC lib/util/md5.o 00:23:24.742 LIB libspdk_util.a 00:23:24.742 SO libspdk_util.so.10.1 00:23:24.742 LIB libspdk_trace_parser.a 00:23:24.742 SO libspdk_trace_parser.so.6.0 00:23:24.742 SYMLINK libspdk_util.so 00:23:24.742 SYMLINK libspdk_trace_parser.so 00:23:24.742 CC lib/env_dpdk/env.o 00:23:24.742 CC lib/env_dpdk/memory.o 00:23:24.742 CC lib/vmd/vmd.o 00:23:24.742 CC lib/conf/conf.o 00:23:24.742 CC lib/vmd/led.o 00:23:24.742 CC lib/json/json_parse.o 00:23:24.742 CC lib/env_dpdk/pci.o 00:23:24.742 CC lib/env_dpdk/init.o 00:23:24.742 CC lib/rdma_utils/rdma_utils.o 00:23:24.742 CC lib/idxd/idxd.o 00:23:24.742 CC lib/idxd/idxd_user.o 00:23:24.742 LIB libspdk_conf.a 00:23:24.742 SO libspdk_conf.so.6.0 00:23:25.002 SYMLINK libspdk_conf.so 00:23:25.002 CC lib/idxd/idxd_kernel.o 00:23:25.002 CC lib/json/json_util.o 00:23:25.002 CC lib/env_dpdk/threads.o 00:23:25.002 CC lib/json/json_write.o 00:23:25.002 CC lib/env_dpdk/pci_ioat.o 00:23:25.002 CC lib/env_dpdk/pci_virtio.o 00:23:25.002 LIB libspdk_rdma_utils.a 00:23:25.002 SO libspdk_rdma_utils.so.1.0 00:23:25.002 CC lib/env_dpdk/pci_vmd.o 00:23:25.260 CC lib/env_dpdk/pci_idxd.o 00:23:25.260 SYMLINK libspdk_rdma_utils.so 00:23:25.260 CC lib/env_dpdk/pci_event.o 00:23:25.260 CC lib/env_dpdk/sigbus_handler.o 00:23:25.260 CC lib/env_dpdk/pci_dpdk.o 00:23:25.260 CC lib/env_dpdk/pci_dpdk_2207.o 00:23:25.260 CC lib/env_dpdk/pci_dpdk_2211.o 00:23:25.260 LIB libspdk_json.a 00:23:25.260 LIB libspdk_vmd.a 00:23:25.260 SO libspdk_json.so.6.0 00:23:25.260 LIB libspdk_idxd.a 00:23:25.260 SO libspdk_vmd.so.6.0 00:23:25.260 SO libspdk_idxd.so.12.1 00:23:25.260 SYMLINK libspdk_json.so 00:23:25.518 CC lib/rdma_provider/common.o 00:23:25.518 CC lib/rdma_provider/rdma_provider_verbs.o 00:23:25.518 SYMLINK libspdk_vmd.so 00:23:25.518 SYMLINK libspdk_idxd.so 00:23:25.518 CC lib/jsonrpc/jsonrpc_server.o 00:23:25.518 CC lib/jsonrpc/jsonrpc_client.o 00:23:25.518 CC lib/jsonrpc/jsonrpc_server_tcp.o 00:23:25.518 CC lib/jsonrpc/jsonrpc_client_tcp.o 00:23:25.518 LIB libspdk_rdma_provider.a 00:23:25.518 SO libspdk_rdma_provider.so.7.0 00:23:25.776 SYMLINK libspdk_rdma_provider.so 00:23:25.776 LIB libspdk_jsonrpc.a 00:23:25.776 SO libspdk_jsonrpc.so.6.0 00:23:26.035 SYMLINK libspdk_jsonrpc.so 00:23:26.035 LIB libspdk_env_dpdk.a 00:23:26.035 CC lib/rpc/rpc.o 00:23:26.292 SO libspdk_env_dpdk.so.15.1 00:23:26.292 SYMLINK libspdk_env_dpdk.so 00:23:26.292 LIB libspdk_rpc.a 00:23:26.292 SO libspdk_rpc.so.6.0 00:23:26.550 SYMLINK libspdk_rpc.so 00:23:26.550 CC lib/trace/trace.o 00:23:26.550 CC lib/trace/trace_flags.o 00:23:26.550 CC lib/trace/trace_rpc.o 00:23:26.550 CC lib/keyring/keyring.o 00:23:26.550 CC lib/keyring/keyring_rpc.o 00:23:26.551 CC lib/notify/notify.o 00:23:26.551 CC lib/notify/notify_rpc.o 00:23:26.808 LIB libspdk_notify.a 00:23:26.808 SO libspdk_notify.so.6.0 00:23:26.808 LIB libspdk_keyring.a 00:23:26.808 SYMLINK libspdk_notify.so 00:23:26.808 LIB libspdk_trace.a 00:23:26.808 SO libspdk_keyring.so.2.0 00:23:26.808 SO libspdk_trace.so.11.0 00:23:26.808 SYMLINK libspdk_keyring.so 00:23:27.068 SYMLINK libspdk_trace.so 00:23:27.068 CC lib/thread/thread.o 00:23:27.068 CC lib/thread/iobuf.o 00:23:27.068 CC lib/sock/sock.o 00:23:27.068 CC lib/sock/sock_rpc.o 00:23:27.644 LIB libspdk_sock.a 00:23:27.644 SO libspdk_sock.so.10.0 00:23:27.644 SYMLINK libspdk_sock.so 00:23:27.901 CC lib/nvme/nvme_ctrlr_cmd.o 00:23:27.901 CC lib/nvme/nvme_fabric.o 00:23:27.901 CC lib/nvme/nvme_ctrlr.o 00:23:27.901 CC lib/nvme/nvme_ns_cmd.o 00:23:27.901 CC lib/nvme/nvme_ns.o 00:23:27.901 CC lib/nvme/nvme_pcie.o 00:23:27.901 CC lib/nvme/nvme_pcie_common.o 00:23:27.901 CC lib/nvme/nvme_qpair.o 00:23:27.901 CC lib/nvme/nvme.o 00:23:28.466 CC lib/nvme/nvme_quirks.o 00:23:28.466 CC lib/nvme/nvme_transport.o 00:23:28.724 CC lib/nvme/nvme_discovery.o 00:23:28.724 CC lib/nvme/nvme_ctrlr_ocssd_cmd.o 00:23:28.724 CC lib/nvme/nvme_ns_ocssd_cmd.o 00:23:28.724 LIB libspdk_thread.a 00:23:28.724 CC lib/nvme/nvme_tcp.o 00:23:28.724 SO libspdk_thread.so.11.0 00:23:28.724 SYMLINK libspdk_thread.so 00:23:28.724 CC lib/nvme/nvme_opal.o 00:23:28.982 CC lib/nvme/nvme_io_msg.o 00:23:28.982 CC lib/nvme/nvme_poll_group.o 00:23:28.982 CC lib/accel/accel.o 00:23:29.240 CC lib/nvme/nvme_zns.o 00:23:29.240 CC lib/nvme/nvme_stubs.o 00:23:29.240 CC lib/nvme/nvme_auth.o 00:23:29.497 CC lib/accel/accel_rpc.o 00:23:29.497 CC lib/blob/blobstore.o 00:23:29.497 CC lib/init/json_config.o 00:23:29.755 CC lib/nvme/nvme_cuse.o 00:23:29.756 CC lib/nvme/nvme_rdma.o 00:23:29.756 CC lib/init/subsystem.o 00:23:29.756 CC lib/virtio/virtio.o 00:23:30.013 CC lib/fsdev/fsdev.o 00:23:30.013 CC lib/init/subsystem_rpc.o 00:23:30.271 CC lib/init/rpc.o 00:23:30.271 CC lib/virtio/virtio_vhost_user.o 00:23:30.271 CC lib/virtio/virtio_vfio_user.o 00:23:30.271 CC lib/virtio/virtio_pci.o 00:23:30.271 LIB libspdk_init.a 00:23:30.271 SO libspdk_init.so.6.0 00:23:30.271 CC lib/blob/request.o 00:23:30.529 SYMLINK libspdk_init.so 00:23:30.529 CC lib/blob/zeroes.o 00:23:30.529 CC lib/blob/blob_bs_dev.o 00:23:30.529 CC lib/accel/accel_sw.o 00:23:30.529 LIB libspdk_virtio.a 00:23:30.529 SO libspdk_virtio.so.7.0 00:23:30.529 CC lib/fsdev/fsdev_io.o 00:23:30.529 CC lib/fsdev/fsdev_rpc.o 00:23:30.529 SYMLINK libspdk_virtio.so 00:23:30.787 CC lib/event/log_rpc.o 00:23:30.787 CC lib/event/app.o 00:23:30.787 CC lib/event/app_rpc.o 00:23:30.787 CC lib/event/reactor.o 00:23:30.787 CC lib/event/scheduler_static.o 00:23:30.787 LIB libspdk_accel.a 00:23:30.787 SO libspdk_accel.so.16.0 00:23:31.044 SYMLINK libspdk_accel.so 00:23:31.044 LIB libspdk_fsdev.a 00:23:31.044 SO libspdk_fsdev.so.2.0 00:23:31.044 SYMLINK libspdk_fsdev.so 00:23:31.044 LIB libspdk_nvme.a 00:23:31.044 CC lib/bdev/bdev_rpc.o 00:23:31.044 CC lib/bdev/bdev.o 00:23:31.044 CC lib/bdev/bdev_zone.o 00:23:31.044 CC lib/bdev/scsi_nvme.o 00:23:31.044 CC lib/bdev/part.o 00:23:31.044 LIB libspdk_event.a 00:23:31.302 SO libspdk_event.so.14.0 00:23:31.302 CC lib/fuse_dispatcher/fuse_dispatcher.o 00:23:31.302 SYMLINK libspdk_event.so 00:23:31.302 SO libspdk_nvme.so.15.0 00:23:31.561 SYMLINK libspdk_nvme.so 00:23:31.819 LIB libspdk_fuse_dispatcher.a 00:23:31.819 SO libspdk_fuse_dispatcher.so.1.0 00:23:32.076 SYMLINK libspdk_fuse_dispatcher.so 00:23:33.010 LIB libspdk_blob.a 00:23:33.268 SO libspdk_blob.so.11.0 00:23:33.268 SYMLINK libspdk_blob.so 00:23:33.526 CC lib/lvol/lvol.o 00:23:33.526 CC lib/blobfs/blobfs.o 00:23:33.526 CC lib/blobfs/tree.o 00:23:34.092 LIB libspdk_bdev.a 00:23:34.092 SO libspdk_bdev.so.17.0 00:23:34.350 SYMLINK libspdk_bdev.so 00:23:34.350 LIB libspdk_blobfs.a 00:23:34.350 SO libspdk_blobfs.so.10.0 00:23:34.350 CC lib/ublk/ublk.o 00:23:34.350 CC lib/nvmf/ctrlr.o 00:23:34.350 CC lib/ublk/ublk_rpc.o 00:23:34.350 CC lib/nvmf/ctrlr_discovery.o 00:23:34.350 CC lib/nvmf/ctrlr_bdev.o 00:23:34.350 CC lib/scsi/dev.o 00:23:34.350 CC lib/nbd/nbd.o 00:23:34.350 CC lib/ftl/ftl_core.o 00:23:34.350 SYMLINK libspdk_blobfs.so 00:23:34.350 CC lib/scsi/lun.o 00:23:34.350 LIB libspdk_lvol.a 00:23:34.350 SO libspdk_lvol.so.10.0 00:23:34.607 SYMLINK libspdk_lvol.so 00:23:34.607 CC lib/scsi/port.o 00:23:34.607 CC lib/scsi/scsi.o 00:23:34.607 CC lib/nvmf/subsystem.o 00:23:34.607 CC lib/scsi/scsi_bdev.o 00:23:34.607 CC lib/scsi/scsi_pr.o 00:23:34.865 CC lib/nvmf/nvmf.o 00:23:34.865 CC lib/scsi/scsi_rpc.o 00:23:34.865 CC lib/ftl/ftl_init.o 00:23:34.865 CC lib/nbd/nbd_rpc.o 00:23:34.865 CC lib/scsi/task.o 00:23:34.865 LIB libspdk_nbd.a 00:23:34.865 CC lib/ftl/ftl_layout.o 00:23:35.123 SO libspdk_nbd.so.7.0 00:23:35.123 SYMLINK libspdk_nbd.so 00:23:35.123 CC lib/ftl/ftl_debug.o 00:23:35.123 CC lib/ftl/ftl_io.o 00:23:35.123 CC lib/ftl/ftl_sb.o 00:23:35.123 LIB libspdk_ublk.a 00:23:35.123 SO libspdk_ublk.so.3.0 00:23:35.123 CC lib/nvmf/nvmf_rpc.o 00:23:35.123 SYMLINK libspdk_ublk.so 00:23:35.123 LIB libspdk_scsi.a 00:23:35.123 CC lib/nvmf/transport.o 00:23:35.123 CC lib/ftl/ftl_l2p.o 00:23:35.123 SO libspdk_scsi.so.9.0 00:23:35.380 CC lib/ftl/ftl_l2p_flat.o 00:23:35.380 SYMLINK libspdk_scsi.so 00:23:35.381 CC lib/nvmf/tcp.o 00:23:35.381 CC lib/nvmf/stubs.o 00:23:35.381 CC lib/nvmf/mdns_server.o 00:23:35.381 CC lib/ftl/ftl_nv_cache.o 00:23:35.381 CC lib/ftl/ftl_band.o 00:23:35.638 CC lib/nvmf/rdma.o 00:23:35.638 CC lib/nvmf/auth.o 00:23:35.896 CC lib/iscsi/conn.o 00:23:35.896 CC lib/iscsi/init_grp.o 00:23:35.896 CC lib/ftl/ftl_band_ops.o 00:23:35.896 CC lib/ftl/ftl_writer.o 00:23:35.896 CC lib/vhost/vhost.o 00:23:36.153 CC lib/iscsi/iscsi.o 00:23:36.153 CC lib/iscsi/param.o 00:23:36.153 CC lib/vhost/vhost_rpc.o 00:23:36.153 CC lib/ftl/ftl_rq.o 00:23:36.410 CC lib/vhost/vhost_scsi.o 00:23:36.410 CC lib/ftl/ftl_reloc.o 00:23:36.410 CC lib/iscsi/portal_grp.o 00:23:36.410 CC lib/iscsi/tgt_node.o 00:23:36.410 CC lib/iscsi/iscsi_subsystem.o 00:23:36.667 CC lib/vhost/vhost_blk.o 00:23:36.667 CC lib/ftl/ftl_l2p_cache.o 00:23:36.667 CC lib/vhost/rte_vhost_user.o 00:23:36.667 CC lib/ftl/ftl_p2l.o 00:23:36.925 CC lib/iscsi/iscsi_rpc.o 00:23:36.925 CC lib/iscsi/task.o 00:23:36.925 CC lib/ftl/ftl_p2l_log.o 00:23:37.183 CC lib/ftl/mngt/ftl_mngt.o 00:23:37.183 CC lib/ftl/mngt/ftl_mngt_bdev.o 00:23:37.183 CC lib/ftl/mngt/ftl_mngt_shutdown.o 00:23:37.183 CC lib/ftl/mngt/ftl_mngt_startup.o 00:23:37.183 CC lib/ftl/mngt/ftl_mngt_md.o 00:23:37.440 CC lib/ftl/mngt/ftl_mngt_misc.o 00:23:37.440 CC lib/ftl/mngt/ftl_mngt_ioch.o 00:23:37.441 CC lib/ftl/mngt/ftl_mngt_l2p.o 00:23:37.441 CC lib/ftl/mngt/ftl_mngt_band.o 00:23:37.441 CC lib/ftl/mngt/ftl_mngt_self_test.o 00:23:37.441 LIB libspdk_iscsi.a 00:23:37.441 CC lib/ftl/mngt/ftl_mngt_p2l.o 00:23:37.441 CC lib/ftl/mngt/ftl_mngt_recovery.o 00:23:37.441 CC lib/ftl/mngt/ftl_mngt_upgrade.o 00:23:37.441 CC lib/ftl/utils/ftl_conf.o 00:23:37.441 LIB libspdk_vhost.a 00:23:37.441 SO libspdk_iscsi.so.8.0 00:23:37.699 CC lib/ftl/utils/ftl_md.o 00:23:37.699 SO libspdk_vhost.so.8.0 00:23:37.699 CC lib/ftl/utils/ftl_mempool.o 00:23:37.699 SYMLINK libspdk_iscsi.so 00:23:37.699 SYMLINK libspdk_vhost.so 00:23:37.699 CC lib/ftl/utils/ftl_property.o 00:23:37.699 CC lib/ftl/utils/ftl_bitmap.o 00:23:37.699 CC lib/ftl/utils/ftl_layout_tracker_bdev.o 00:23:37.699 CC lib/ftl/upgrade/ftl_layout_upgrade.o 00:23:37.699 CC lib/ftl/upgrade/ftl_sb_upgrade.o 00:23:37.699 LIB libspdk_nvmf.a 00:23:37.699 CC lib/ftl/upgrade/ftl_p2l_upgrade.o 00:23:37.957 SO libspdk_nvmf.so.20.0 00:23:37.957 CC lib/ftl/upgrade/ftl_band_upgrade.o 00:23:37.957 CC lib/ftl/upgrade/ftl_chunk_upgrade.o 00:23:37.957 CC lib/ftl/upgrade/ftl_trim_upgrade.o 00:23:37.957 CC lib/ftl/upgrade/ftl_sb_v3.o 00:23:37.957 CC lib/ftl/upgrade/ftl_sb_v5.o 00:23:37.957 CC lib/ftl/nvc/ftl_nvc_dev.o 00:23:37.957 CC lib/ftl/nvc/ftl_nvc_bdev_vss.o 00:23:37.957 CC lib/ftl/nvc/ftl_nvc_bdev_non_vss.o 00:23:37.957 CC lib/ftl/nvc/ftl_nvc_bdev_common.o 00:23:37.957 SYMLINK libspdk_nvmf.so 00:23:37.957 CC lib/ftl/base/ftl_base_dev.o 00:23:37.957 CC lib/ftl/base/ftl_base_bdev.o 00:23:37.957 CC lib/ftl/ftl_trace.o 00:23:38.215 LIB libspdk_ftl.a 00:23:38.473 SO libspdk_ftl.so.9.0 00:23:38.730 SYMLINK libspdk_ftl.so 00:23:38.987 CC module/env_dpdk/env_dpdk_rpc.o 00:23:38.987 CC module/sock/posix/posix.o 00:23:39.245 CC module/fsdev/aio/fsdev_aio.o 00:23:39.245 CC module/scheduler/dpdk_governor/dpdk_governor.o 00:23:39.245 CC module/accel/error/accel_error.o 00:23:39.245 CC module/scheduler/dynamic/scheduler_dynamic.o 00:23:39.245 CC module/accel/ioat/accel_ioat.o 00:23:39.245 CC module/blob/bdev/blob_bdev.o 00:23:39.245 CC module/scheduler/gscheduler/gscheduler.o 00:23:39.245 CC module/keyring/file/keyring.o 00:23:39.245 LIB libspdk_env_dpdk_rpc.a 00:23:39.245 SO libspdk_env_dpdk_rpc.so.6.0 00:23:39.245 SYMLINK libspdk_env_dpdk_rpc.so 00:23:39.245 CC module/accel/error/accel_error_rpc.o 00:23:39.245 LIB libspdk_scheduler_dpdk_governor.a 00:23:39.245 LIB libspdk_scheduler_gscheduler.a 00:23:39.245 CC module/keyring/file/keyring_rpc.o 00:23:39.245 SO libspdk_scheduler_dpdk_governor.so.4.0 00:23:39.245 SO libspdk_scheduler_gscheduler.so.4.0 00:23:39.245 CC module/accel/ioat/accel_ioat_rpc.o 00:23:39.245 LIB libspdk_scheduler_dynamic.a 00:23:39.245 SYMLINK libspdk_scheduler_dpdk_governor.so 00:23:39.245 CC module/fsdev/aio/fsdev_aio_rpc.o 00:23:39.245 SO libspdk_scheduler_dynamic.so.4.0 00:23:39.245 SYMLINK libspdk_scheduler_gscheduler.so 00:23:39.245 CC module/fsdev/aio/linux_aio_mgr.o 00:23:39.245 SYMLINK libspdk_scheduler_dynamic.so 00:23:39.503 LIB libspdk_accel_error.a 00:23:39.503 LIB libspdk_blob_bdev.a 00:23:39.503 LIB libspdk_accel_ioat.a 00:23:39.503 SO libspdk_accel_error.so.2.0 00:23:39.503 SO libspdk_blob_bdev.so.11.0 00:23:39.503 LIB libspdk_keyring_file.a 00:23:39.503 SO libspdk_accel_ioat.so.6.0 00:23:39.503 SO libspdk_keyring_file.so.2.0 00:23:39.503 SYMLINK libspdk_accel_error.so 00:23:39.503 CC module/keyring/linux/keyring.o 00:23:39.503 SYMLINK libspdk_blob_bdev.so 00:23:39.503 SYMLINK libspdk_keyring_file.so 00:23:39.503 SYMLINK libspdk_accel_ioat.so 00:23:39.503 CC module/keyring/linux/keyring_rpc.o 00:23:39.503 CC module/accel/dsa/accel_dsa.o 00:23:39.503 CC module/accel/dsa/accel_dsa_rpc.o 00:23:39.503 CC module/accel/iaa/accel_iaa.o 00:23:39.761 LIB libspdk_keyring_linux.a 00:23:39.761 SO libspdk_keyring_linux.so.1.0 00:23:39.761 CC module/accel/iaa/accel_iaa_rpc.o 00:23:39.761 CC module/bdev/error/vbdev_error.o 00:23:39.761 CC module/bdev/delay/vbdev_delay.o 00:23:39.761 SYMLINK libspdk_keyring_linux.so 00:23:39.761 CC module/bdev/error/vbdev_error_rpc.o 00:23:39.761 CC module/blobfs/bdev/blobfs_bdev.o 00:23:39.761 LIB libspdk_fsdev_aio.a 00:23:39.761 CC module/bdev/gpt/gpt.o 00:23:39.761 SO libspdk_fsdev_aio.so.1.0 00:23:39.761 LIB libspdk_accel_iaa.a 00:23:39.761 LIB libspdk_accel_dsa.a 00:23:39.761 SO libspdk_accel_dsa.so.5.0 00:23:39.761 SO libspdk_accel_iaa.so.3.0 00:23:39.761 LIB libspdk_sock_posix.a 00:23:40.019 CC module/bdev/gpt/vbdev_gpt.o 00:23:40.019 SO libspdk_sock_posix.so.6.0 00:23:40.019 SYMLINK libspdk_fsdev_aio.so 00:23:40.019 CC module/blobfs/bdev/blobfs_bdev_rpc.o 00:23:40.019 CC module/bdev/delay/vbdev_delay_rpc.o 00:23:40.019 SYMLINK libspdk_accel_dsa.so 00:23:40.019 SYMLINK libspdk_accel_iaa.so 00:23:40.019 CC module/bdev/lvol/vbdev_lvol.o 00:23:40.019 LIB libspdk_bdev_error.a 00:23:40.019 SYMLINK libspdk_sock_posix.so 00:23:40.019 SO libspdk_bdev_error.so.6.0 00:23:40.019 SYMLINK libspdk_bdev_error.so 00:23:40.019 CC module/bdev/malloc/bdev_malloc.o 00:23:40.019 CC module/bdev/lvol/vbdev_lvol_rpc.o 00:23:40.019 LIB libspdk_blobfs_bdev.a 00:23:40.019 CC module/bdev/null/bdev_null.o 00:23:40.019 LIB libspdk_bdev_delay.a 00:23:40.019 SO libspdk_blobfs_bdev.so.6.0 00:23:40.019 SO libspdk_bdev_delay.so.6.0 00:23:40.019 CC module/bdev/nvme/bdev_nvme.o 00:23:40.019 CC module/bdev/passthru/vbdev_passthru.o 00:23:40.278 SYMLINK libspdk_blobfs_bdev.so 00:23:40.278 LIB libspdk_bdev_gpt.a 00:23:40.278 CC module/bdev/malloc/bdev_malloc_rpc.o 00:23:40.278 SYMLINK libspdk_bdev_delay.so 00:23:40.278 CC module/bdev/null/bdev_null_rpc.o 00:23:40.278 SO libspdk_bdev_gpt.so.6.0 00:23:40.278 CC module/bdev/raid/bdev_raid.o 00:23:40.278 SYMLINK libspdk_bdev_gpt.so 00:23:40.278 CC module/bdev/nvme/bdev_nvme_rpc.o 00:23:40.278 CC module/bdev/nvme/nvme_rpc.o 00:23:40.278 LIB libspdk_bdev_null.a 00:23:40.278 SO libspdk_bdev_null.so.6.0 00:23:40.278 CC module/bdev/split/vbdev_split.o 00:23:40.535 SYMLINK libspdk_bdev_null.so 00:23:40.535 CC module/bdev/nvme/bdev_mdns_client.o 00:23:40.535 CC module/bdev/passthru/vbdev_passthru_rpc.o 00:23:40.535 LIB libspdk_bdev_malloc.a 00:23:40.535 LIB libspdk_bdev_lvol.a 00:23:40.535 SO libspdk_bdev_malloc.so.6.0 00:23:40.535 SO libspdk_bdev_lvol.so.6.0 00:23:40.535 SYMLINK libspdk_bdev_malloc.so 00:23:40.535 CC module/bdev/nvme/vbdev_opal.o 00:23:40.535 CC module/bdev/nvme/vbdev_opal_rpc.o 00:23:40.535 SYMLINK libspdk_bdev_lvol.so 00:23:40.535 CC module/bdev/zone_block/vbdev_zone_block.o 00:23:40.535 CC module/bdev/zone_block/vbdev_zone_block_rpc.o 00:23:40.535 CC module/bdev/split/vbdev_split_rpc.o 00:23:40.535 LIB libspdk_bdev_passthru.a 00:23:40.535 SO libspdk_bdev_passthru.so.6.0 00:23:40.792 CC module/bdev/aio/bdev_aio.o 00:23:40.792 SYMLINK libspdk_bdev_passthru.so 00:23:40.793 CC module/bdev/aio/bdev_aio_rpc.o 00:23:40.793 CC module/bdev/nvme/bdev_nvme_cuse_rpc.o 00:23:40.793 LIB libspdk_bdev_split.a 00:23:40.793 SO libspdk_bdev_split.so.6.0 00:23:40.793 SYMLINK libspdk_bdev_split.so 00:23:40.793 CC module/bdev/raid/bdev_raid_rpc.o 00:23:40.793 CC module/bdev/raid/bdev_raid_sb.o 00:23:40.793 LIB libspdk_bdev_zone_block.a 00:23:41.051 CC module/bdev/ftl/bdev_ftl.o 00:23:41.052 SO libspdk_bdev_zone_block.so.6.0 00:23:41.052 CC module/bdev/iscsi/bdev_iscsi.o 00:23:41.052 CC module/bdev/virtio/bdev_virtio_scsi.o 00:23:41.052 CC module/bdev/raid/raid0.o 00:23:41.052 SYMLINK libspdk_bdev_zone_block.so 00:23:41.052 CC module/bdev/iscsi/bdev_iscsi_rpc.o 00:23:41.052 LIB libspdk_bdev_aio.a 00:23:41.052 SO libspdk_bdev_aio.so.6.0 00:23:41.052 CC module/bdev/ftl/bdev_ftl_rpc.o 00:23:41.052 SYMLINK libspdk_bdev_aio.so 00:23:41.052 CC module/bdev/virtio/bdev_virtio_blk.o 00:23:41.052 CC module/bdev/raid/raid1.o 00:23:41.052 CC module/bdev/virtio/bdev_virtio_rpc.o 00:23:41.309 CC module/bdev/raid/concat.o 00:23:41.310 CC module/bdev/raid/raid5f.o 00:23:41.310 LIB libspdk_bdev_ftl.a 00:23:41.310 SO libspdk_bdev_ftl.so.6.0 00:23:41.310 LIB libspdk_bdev_iscsi.a 00:23:41.310 SO libspdk_bdev_iscsi.so.6.0 00:23:41.310 SYMLINK libspdk_bdev_ftl.so 00:23:41.310 SYMLINK libspdk_bdev_iscsi.so 00:23:41.568 LIB libspdk_bdev_virtio.a 00:23:41.568 SO libspdk_bdev_virtio.so.6.0 00:23:41.568 SYMLINK libspdk_bdev_virtio.so 00:23:41.826 LIB libspdk_bdev_raid.a 00:23:41.826 SO libspdk_bdev_raid.so.6.0 00:23:41.826 SYMLINK libspdk_bdev_raid.so 00:23:42.758 LIB libspdk_bdev_nvme.a 00:23:42.758 SO libspdk_bdev_nvme.so.7.1 00:23:42.758 SYMLINK libspdk_bdev_nvme.so 00:23:43.323 CC module/event/subsystems/scheduler/scheduler.o 00:23:43.323 CC module/event/subsystems/sock/sock.o 00:23:43.323 CC module/event/subsystems/keyring/keyring.o 00:23:43.323 CC module/event/subsystems/fsdev/fsdev.o 00:23:43.323 CC module/event/subsystems/vmd/vmd.o 00:23:43.323 CC module/event/subsystems/vmd/vmd_rpc.o 00:23:43.323 CC module/event/subsystems/vhost_blk/vhost_blk.o 00:23:43.323 CC module/event/subsystems/iobuf/iobuf.o 00:23:43.323 CC module/event/subsystems/iobuf/iobuf_rpc.o 00:23:43.323 LIB libspdk_event_keyring.a 00:23:43.323 LIB libspdk_event_fsdev.a 00:23:43.323 LIB libspdk_event_scheduler.a 00:23:43.323 LIB libspdk_event_vhost_blk.a 00:23:43.323 LIB libspdk_event_sock.a 00:23:43.323 LIB libspdk_event_vmd.a 00:23:43.323 SO libspdk_event_scheduler.so.4.0 00:23:43.323 SO libspdk_event_fsdev.so.1.0 00:23:43.323 SO libspdk_event_sock.so.5.0 00:23:43.323 SO libspdk_event_vhost_blk.so.3.0 00:23:43.323 SO libspdk_event_keyring.so.1.0 00:23:43.323 SO libspdk_event_vmd.so.6.0 00:23:43.323 LIB libspdk_event_iobuf.a 00:23:43.323 SYMLINK libspdk_event_sock.so 00:23:43.323 SYMLINK libspdk_event_vhost_blk.so 00:23:43.323 SYMLINK libspdk_event_scheduler.so 00:23:43.323 SYMLINK libspdk_event_keyring.so 00:23:43.323 SYMLINK libspdk_event_fsdev.so 00:23:43.323 SO libspdk_event_iobuf.so.3.0 00:23:43.323 SYMLINK libspdk_event_vmd.so 00:23:43.582 SYMLINK libspdk_event_iobuf.so 00:23:43.582 CC module/event/subsystems/accel/accel.o 00:23:43.840 LIB libspdk_event_accel.a 00:23:43.840 SO libspdk_event_accel.so.6.0 00:23:43.840 SYMLINK libspdk_event_accel.so 00:23:44.098 CC module/event/subsystems/bdev/bdev.o 00:23:44.397 LIB libspdk_event_bdev.a 00:23:44.397 SO libspdk_event_bdev.so.6.0 00:23:44.656 SYMLINK libspdk_event_bdev.so 00:23:44.656 CC module/event/subsystems/nbd/nbd.o 00:23:44.656 CC module/event/subsystems/scsi/scsi.o 00:23:44.656 CC module/event/subsystems/ublk/ublk.o 00:23:44.656 CC module/event/subsystems/nvmf/nvmf_tgt.o 00:23:44.656 CC module/event/subsystems/nvmf/nvmf_rpc.o 00:23:44.656 LIB libspdk_event_nbd.a 00:23:44.656 LIB libspdk_event_scsi.a 00:23:44.656 LIB libspdk_event_ublk.a 00:23:44.656 SO libspdk_event_nbd.so.6.0 00:23:44.656 SO libspdk_event_scsi.so.6.0 00:23:44.656 SO libspdk_event_ublk.so.3.0 00:23:44.656 SYMLINK libspdk_event_nbd.so 00:23:44.656 SYMLINK libspdk_event_ublk.so 00:23:44.656 SYMLINK libspdk_event_scsi.so 00:23:44.656 LIB libspdk_event_nvmf.a 00:23:44.915 SO libspdk_event_nvmf.so.6.0 00:23:44.915 SYMLINK libspdk_event_nvmf.so 00:23:44.915 CC module/event/subsystems/iscsi/iscsi.o 00:23:44.915 CC module/event/subsystems/vhost_scsi/vhost_scsi.o 00:23:45.173 LIB libspdk_event_iscsi.a 00:23:45.173 LIB libspdk_event_vhost_scsi.a 00:23:45.173 SO libspdk_event_vhost_scsi.so.3.0 00:23:45.173 SO libspdk_event_iscsi.so.6.0 00:23:45.173 SYMLINK libspdk_event_vhost_scsi.so 00:23:45.173 SYMLINK libspdk_event_iscsi.so 00:23:45.173 SO libspdk.so.6.0 00:23:45.173 SYMLINK libspdk.so 00:23:45.431 CC app/trace_record/trace_record.o 00:23:45.431 CXX app/trace/trace.o 00:23:45.431 CC app/spdk_lspci/spdk_lspci.o 00:23:45.431 CC examples/interrupt_tgt/interrupt_tgt.o 00:23:45.431 CC app/nvmf_tgt/nvmf_main.o 00:23:45.431 CC app/iscsi_tgt/iscsi_tgt.o 00:23:45.431 CC app/spdk_tgt/spdk_tgt.o 00:23:45.689 CC examples/util/zipf/zipf.o 00:23:45.689 CC examples/ioat/perf/perf.o 00:23:45.689 CC test/thread/poller_perf/poller_perf.o 00:23:45.689 LINK spdk_lspci 00:23:45.689 LINK interrupt_tgt 00:23:45.689 LINK zipf 00:23:45.689 LINK spdk_trace_record 00:23:45.689 LINK iscsi_tgt 00:23:45.689 LINK poller_perf 00:23:45.689 LINK nvmf_tgt 00:23:45.689 LINK spdk_tgt 00:23:45.689 LINK ioat_perf 00:23:45.948 LINK spdk_trace 00:23:45.948 CC app/spdk_nvme_perf/perf.o 00:23:45.948 TEST_HEADER include/spdk/accel.h 00:23:45.948 TEST_HEADER include/spdk/accel_module.h 00:23:45.948 TEST_HEADER include/spdk/assert.h 00:23:45.948 TEST_HEADER include/spdk/barrier.h 00:23:45.948 TEST_HEADER include/spdk/base64.h 00:23:45.948 TEST_HEADER include/spdk/bdev.h 00:23:45.948 CC examples/ioat/verify/verify.o 00:23:45.948 TEST_HEADER include/spdk/bdev_module.h 00:23:45.948 TEST_HEADER include/spdk/bdev_zone.h 00:23:45.948 TEST_HEADER include/spdk/bit_array.h 00:23:45.948 TEST_HEADER include/spdk/bit_pool.h 00:23:45.948 TEST_HEADER include/spdk/blob_bdev.h 00:23:45.948 TEST_HEADER include/spdk/blobfs_bdev.h 00:23:45.948 TEST_HEADER include/spdk/blobfs.h 00:23:45.948 TEST_HEADER include/spdk/blob.h 00:23:45.948 TEST_HEADER include/spdk/conf.h 00:23:45.948 TEST_HEADER include/spdk/config.h 00:23:45.948 TEST_HEADER include/spdk/cpuset.h 00:23:45.948 TEST_HEADER include/spdk/crc16.h 00:23:45.948 TEST_HEADER include/spdk/crc32.h 00:23:45.948 TEST_HEADER include/spdk/crc64.h 00:23:45.948 TEST_HEADER include/spdk/dif.h 00:23:45.948 TEST_HEADER include/spdk/dma.h 00:23:45.948 TEST_HEADER include/spdk/endian.h 00:23:45.948 TEST_HEADER include/spdk/env_dpdk.h 00:23:45.948 TEST_HEADER include/spdk/env.h 00:23:45.948 TEST_HEADER include/spdk/event.h 00:23:45.948 TEST_HEADER include/spdk/fd_group.h 00:23:45.948 TEST_HEADER include/spdk/fd.h 00:23:45.948 TEST_HEADER include/spdk/file.h 00:23:45.948 TEST_HEADER include/spdk/fsdev.h 00:23:45.948 TEST_HEADER include/spdk/fsdev_module.h 00:23:45.948 TEST_HEADER include/spdk/ftl.h 00:23:45.948 TEST_HEADER include/spdk/fuse_dispatcher.h 00:23:45.948 TEST_HEADER include/spdk/gpt_spec.h 00:23:45.948 TEST_HEADER include/spdk/hexlify.h 00:23:45.948 TEST_HEADER include/spdk/histogram_data.h 00:23:45.948 TEST_HEADER include/spdk/idxd.h 00:23:45.948 CC app/spdk_nvme_identify/identify.o 00:23:45.948 TEST_HEADER include/spdk/idxd_spec.h 00:23:45.948 TEST_HEADER include/spdk/init.h 00:23:45.948 CC app/spdk_nvme_discover/discovery_aer.o 00:23:45.948 TEST_HEADER include/spdk/ioat.h 00:23:45.948 TEST_HEADER include/spdk/ioat_spec.h 00:23:45.948 TEST_HEADER include/spdk/iscsi_spec.h 00:23:45.948 TEST_HEADER include/spdk/json.h 00:23:45.948 TEST_HEADER include/spdk/jsonrpc.h 00:23:45.948 TEST_HEADER include/spdk/keyring.h 00:23:45.948 TEST_HEADER include/spdk/keyring_module.h 00:23:45.948 CC app/spdk_top/spdk_top.o 00:23:45.948 TEST_HEADER include/spdk/likely.h 00:23:45.948 TEST_HEADER include/spdk/log.h 00:23:45.948 TEST_HEADER include/spdk/lvol.h 00:23:45.948 TEST_HEADER include/spdk/md5.h 00:23:45.948 TEST_HEADER include/spdk/memory.h 00:23:45.948 TEST_HEADER include/spdk/mmio.h 00:23:45.948 TEST_HEADER include/spdk/nbd.h 00:23:45.948 TEST_HEADER include/spdk/net.h 00:23:45.948 TEST_HEADER include/spdk/notify.h 00:23:45.948 TEST_HEADER include/spdk/nvme.h 00:23:45.948 CC test/dma/test_dma/test_dma.o 00:23:45.948 TEST_HEADER include/spdk/nvme_intel.h 00:23:45.948 TEST_HEADER include/spdk/nvme_ocssd.h 00:23:45.948 TEST_HEADER include/spdk/nvme_ocssd_spec.h 00:23:45.948 TEST_HEADER include/spdk/nvme_spec.h 00:23:45.948 TEST_HEADER include/spdk/nvme_zns.h 00:23:45.948 TEST_HEADER include/spdk/nvmf_cmd.h 00:23:45.948 TEST_HEADER include/spdk/nvmf_fc_spec.h 00:23:45.948 TEST_HEADER include/spdk/nvmf.h 00:23:45.948 TEST_HEADER include/spdk/nvmf_spec.h 00:23:45.948 TEST_HEADER include/spdk/nvmf_transport.h 00:23:45.948 TEST_HEADER include/spdk/opal.h 00:23:45.948 TEST_HEADER include/spdk/opal_spec.h 00:23:45.948 TEST_HEADER include/spdk/pci_ids.h 00:23:45.948 TEST_HEADER include/spdk/pipe.h 00:23:45.948 CC test/app/bdev_svc/bdev_svc.o 00:23:45.948 CC app/spdk_dd/spdk_dd.o 00:23:45.948 TEST_HEADER include/spdk/queue.h 00:23:45.948 TEST_HEADER include/spdk/reduce.h 00:23:45.948 TEST_HEADER include/spdk/rpc.h 00:23:45.948 TEST_HEADER include/spdk/scheduler.h 00:23:45.948 TEST_HEADER include/spdk/scsi.h 00:23:45.948 TEST_HEADER include/spdk/scsi_spec.h 00:23:45.948 TEST_HEADER include/spdk/sock.h 00:23:45.948 TEST_HEADER include/spdk/stdinc.h 00:23:45.948 TEST_HEADER include/spdk/string.h 00:23:45.948 TEST_HEADER include/spdk/thread.h 00:23:45.948 TEST_HEADER include/spdk/trace.h 00:23:45.948 TEST_HEADER include/spdk/trace_parser.h 00:23:45.948 TEST_HEADER include/spdk/tree.h 00:23:46.205 TEST_HEADER include/spdk/ublk.h 00:23:46.205 TEST_HEADER include/spdk/util.h 00:23:46.205 TEST_HEADER include/spdk/uuid.h 00:23:46.205 TEST_HEADER include/spdk/version.h 00:23:46.205 TEST_HEADER include/spdk/vfio_user_pci.h 00:23:46.205 TEST_HEADER include/spdk/vfio_user_spec.h 00:23:46.205 TEST_HEADER include/spdk/vhost.h 00:23:46.205 TEST_HEADER include/spdk/vmd.h 00:23:46.205 TEST_HEADER include/spdk/xor.h 00:23:46.205 TEST_HEADER include/spdk/zipf.h 00:23:46.205 CXX test/cpp_headers/accel.o 00:23:46.205 LINK verify 00:23:46.205 CC app/fio/nvme/fio_plugin.o 00:23:46.205 LINK spdk_nvme_discover 00:23:46.205 LINK bdev_svc 00:23:46.205 CXX test/cpp_headers/accel_module.o 00:23:46.463 CXX test/cpp_headers/assert.o 00:23:46.463 LINK spdk_dd 00:23:46.463 CC app/fio/bdev/fio_plugin.o 00:23:46.463 CC examples/thread/thread/thread_ex.o 00:23:46.463 CC test/app/fuzz/nvme_fuzz/nvme_fuzz.o 00:23:46.463 LINK test_dma 00:23:46.463 CXX test/cpp_headers/barrier.o 00:23:46.463 CXX test/cpp_headers/base64.o 00:23:46.721 LINK spdk_nvme 00:23:46.721 LINK thread 00:23:46.721 CXX test/cpp_headers/bdev.o 00:23:46.721 LINK spdk_nvme_perf 00:23:46.721 CC test/app/fuzz/iscsi_fuzz/iscsi_fuzz.o 00:23:46.721 CC examples/sock/hello_world/hello_sock.o 00:23:46.721 LINK spdk_nvme_identify 00:23:46.721 CC test/app/fuzz/vhost_fuzz/vhost_fuzz_rpc.o 00:23:46.979 CXX test/cpp_headers/bdev_module.o 00:23:46.979 CC test/app/histogram_perf/histogram_perf.o 00:23:46.979 LINK spdk_bdev 00:23:46.979 LINK nvme_fuzz 00:23:46.979 LINK spdk_top 00:23:46.979 CC test/app/fuzz/vhost_fuzz/vhost_fuzz.o 00:23:46.979 CC app/vhost/vhost.o 00:23:46.979 CC test/app/jsoncat/jsoncat.o 00:23:46.979 CXX test/cpp_headers/bdev_zone.o 00:23:46.979 LINK histogram_perf 00:23:46.979 CXX test/cpp_headers/bit_array.o 00:23:46.979 LINK hello_sock 00:23:46.979 CC test/app/stub/stub.o 00:23:46.979 CXX test/cpp_headers/bit_pool.o 00:23:47.237 LINK vhost 00:23:47.237 CXX test/cpp_headers/blob_bdev.o 00:23:47.237 LINK jsoncat 00:23:47.237 CXX test/cpp_headers/blobfs_bdev.o 00:23:47.237 LINK stub 00:23:47.237 CXX test/cpp_headers/blobfs.o 00:23:47.237 CXX test/cpp_headers/blob.o 00:23:47.495 CC examples/vmd/lsvmd/lsvmd.o 00:23:47.495 CC examples/vmd/led/led.o 00:23:47.495 CXX test/cpp_headers/conf.o 00:23:47.495 LINK vhost_fuzz 00:23:47.495 CC examples/idxd/perf/perf.o 00:23:47.495 CC test/env/mem_callbacks/mem_callbacks.o 00:23:47.495 CC test/env/env_dpdk_post_init/env_dpdk_post_init.o 00:23:47.495 CC test/env/vtophys/vtophys.o 00:23:47.495 CXX test/cpp_headers/config.o 00:23:47.495 LINK lsvmd 00:23:47.495 LINK led 00:23:47.495 CXX test/cpp_headers/cpuset.o 00:23:47.495 CC test/env/memory/memory_ut.o 00:23:47.753 LINK vtophys 00:23:47.753 LINK env_dpdk_post_init 00:23:47.753 CXX test/cpp_headers/crc16.o 00:23:47.753 CC test/event/event_perf/event_perf.o 00:23:47.753 CC test/event/reactor/reactor.o 00:23:47.753 CC test/event/reactor_perf/reactor_perf.o 00:23:47.753 LINK idxd_perf 00:23:47.753 CXX test/cpp_headers/crc32.o 00:23:47.753 LINK event_perf 00:23:47.753 CC test/env/pci/pci_ut.o 00:23:47.753 LINK reactor 00:23:47.753 LINK reactor_perf 00:23:48.011 CC test/nvme/aer/aer.o 00:23:48.011 CXX test/cpp_headers/crc64.o 00:23:48.011 CXX test/cpp_headers/dif.o 00:23:48.011 LINK mem_callbacks 00:23:48.011 CXX test/cpp_headers/dma.o 00:23:48.011 CC examples/fsdev/hello_world/hello_fsdev.o 00:23:48.011 CC test/event/app_repeat/app_repeat.o 00:23:48.011 CXX test/cpp_headers/endian.o 00:23:48.011 CXX test/cpp_headers/env_dpdk.o 00:23:48.270 LINK app_repeat 00:23:48.270 CC test/rpc_client/rpc_client_test.o 00:23:48.270 LINK aer 00:23:48.270 LINK pci_ut 00:23:48.270 CC examples/accel/perf/accel_perf.o 00:23:48.270 CXX test/cpp_headers/env.o 00:23:48.270 LINK hello_fsdev 00:23:48.270 LINK rpc_client_test 00:23:48.270 CXX test/cpp_headers/event.o 00:23:48.270 CC test/accel/dif/dif.o 00:23:48.529 CC test/event/scheduler/scheduler.o 00:23:48.529 CC test/nvme/reset/reset.o 00:23:48.529 CXX test/cpp_headers/fd_group.o 00:23:48.529 CXX test/cpp_headers/fd.o 00:23:48.529 CXX test/cpp_headers/file.o 00:23:48.529 LINK iscsi_fuzz 00:23:48.529 CXX test/cpp_headers/fsdev.o 00:23:48.529 LINK scheduler 00:23:48.787 LINK reset 00:23:48.787 LINK memory_ut 00:23:48.787 CC examples/blob/hello_world/hello_blob.o 00:23:48.787 CXX test/cpp_headers/fsdev_module.o 00:23:48.787 CC examples/nvme/hello_world/hello_world.o 00:23:48.787 LINK accel_perf 00:23:48.787 CXX test/cpp_headers/ftl.o 00:23:48.787 CC test/blobfs/mkfs/mkfs.o 00:23:48.787 CXX test/cpp_headers/fuse_dispatcher.o 00:23:48.787 CXX test/cpp_headers/gpt_spec.o 00:23:48.787 CC examples/blob/cli/blobcli.o 00:23:48.787 CC test/nvme/sgl/sgl.o 00:23:48.787 LINK hello_blob 00:23:49.046 CC examples/nvme/reconnect/reconnect.o 00:23:49.046 LINK hello_world 00:23:49.046 LINK mkfs 00:23:49.046 LINK dif 00:23:49.046 CXX test/cpp_headers/hexlify.o 00:23:49.046 CXX test/cpp_headers/histogram_data.o 00:23:49.046 CC examples/nvme/nvme_manage/nvme_manage.o 00:23:49.046 CXX test/cpp_headers/idxd.o 00:23:49.046 LINK sgl 00:23:49.046 CC examples/bdev/hello_world/hello_bdev.o 00:23:49.046 CXX test/cpp_headers/idxd_spec.o 00:23:49.046 CXX test/cpp_headers/init.o 00:23:49.304 CC examples/bdev/bdevperf/bdevperf.o 00:23:49.304 LINK blobcli 00:23:49.304 CXX test/cpp_headers/ioat.o 00:23:49.304 CXX test/cpp_headers/ioat_spec.o 00:23:49.304 LINK reconnect 00:23:49.304 CC test/nvme/e2edp/nvme_dp.o 00:23:49.304 CC test/lvol/esnap/esnap.o 00:23:49.304 LINK hello_bdev 00:23:49.304 CC test/nvme/overhead/overhead.o 00:23:49.304 CXX test/cpp_headers/iscsi_spec.o 00:23:49.563 LINK nvme_manage 00:23:49.563 CC test/nvme/err_injection/err_injection.o 00:23:49.563 CC test/nvme/startup/startup.o 00:23:49.563 CC test/nvme/reserve/reserve.o 00:23:49.563 CXX test/cpp_headers/json.o 00:23:49.563 CC test/nvme/simple_copy/simple_copy.o 00:23:49.563 LINK nvme_dp 00:23:49.563 LINK startup 00:23:49.563 CC examples/nvme/arbitration/arbitration.o 00:23:49.563 LINK err_injection 00:23:49.563 LINK overhead 00:23:49.563 CXX test/cpp_headers/jsonrpc.o 00:23:49.563 LINK reserve 00:23:49.821 CXX test/cpp_headers/keyring.o 00:23:49.821 CXX test/cpp_headers/keyring_module.o 00:23:49.821 LINK simple_copy 00:23:49.821 CXX test/cpp_headers/likely.o 00:23:49.821 CXX test/cpp_headers/log.o 00:23:49.821 LINK bdevperf 00:23:49.821 CXX test/cpp_headers/lvol.o 00:23:49.821 CC test/nvme/connect_stress/connect_stress.o 00:23:49.821 CXX test/cpp_headers/md5.o 00:23:49.821 CXX test/cpp_headers/memory.o 00:23:49.821 CXX test/cpp_headers/mmio.o 00:23:50.078 CC examples/nvme/hotplug/hotplug.o 00:23:50.078 LINK arbitration 00:23:50.078 CC test/bdev/bdevio/bdevio.o 00:23:50.078 CXX test/cpp_headers/nbd.o 00:23:50.078 CXX test/cpp_headers/net.o 00:23:50.078 CXX test/cpp_headers/notify.o 00:23:50.078 LINK connect_stress 00:23:50.078 CXX test/cpp_headers/nvme.o 00:23:50.078 CXX test/cpp_headers/nvme_intel.o 00:23:50.078 CXX test/cpp_headers/nvme_ocssd.o 00:23:50.078 CC test/nvme/boot_partition/boot_partition.o 00:23:50.078 CC examples/nvme/cmb_copy/cmb_copy.o 00:23:50.078 CC test/nvme/compliance/nvme_compliance.o 00:23:50.078 CC examples/nvme/abort/abort.o 00:23:50.078 CXX test/cpp_headers/nvme_ocssd_spec.o 00:23:50.336 LINK hotplug 00:23:50.336 CXX test/cpp_headers/nvme_spec.o 00:23:50.336 LINK boot_partition 00:23:50.336 CC examples/nvme/pmr_persistence/pmr_persistence.o 00:23:50.336 CXX test/cpp_headers/nvme_zns.o 00:23:50.336 LINK bdevio 00:23:50.336 LINK cmb_copy 00:23:50.336 CXX test/cpp_headers/nvmf_cmd.o 00:23:50.336 CC test/nvme/fused_ordering/fused_ordering.o 00:23:50.336 LINK nvme_compliance 00:23:50.594 CC test/nvme/doorbell_aers/doorbell_aers.o 00:23:50.594 LINK pmr_persistence 00:23:50.594 CXX test/cpp_headers/nvmf_fc_spec.o 00:23:50.594 LINK abort 00:23:50.594 CC test/nvme/fdp/fdp.o 00:23:50.594 CXX test/cpp_headers/nvmf.o 00:23:50.594 CC test/nvme/cuse/cuse.o 00:23:50.594 CXX test/cpp_headers/nvmf_spec.o 00:23:50.594 LINK fused_ordering 00:23:50.594 CXX test/cpp_headers/nvmf_transport.o 00:23:50.594 LINK doorbell_aers 00:23:50.594 CXX test/cpp_headers/opal.o 00:23:50.853 CXX test/cpp_headers/opal_spec.o 00:23:50.853 CXX test/cpp_headers/pci_ids.o 00:23:50.853 CXX test/cpp_headers/pipe.o 00:23:50.853 CXX test/cpp_headers/queue.o 00:23:50.853 CC examples/nvmf/nvmf/nvmf.o 00:23:50.853 CXX test/cpp_headers/reduce.o 00:23:50.853 CXX test/cpp_headers/rpc.o 00:23:50.853 CXX test/cpp_headers/scheduler.o 00:23:50.853 LINK fdp 00:23:50.853 CXX test/cpp_headers/scsi.o 00:23:50.853 CXX test/cpp_headers/scsi_spec.o 00:23:50.853 CXX test/cpp_headers/sock.o 00:23:50.853 CXX test/cpp_headers/stdinc.o 00:23:50.853 CXX test/cpp_headers/string.o 00:23:50.853 CXX test/cpp_headers/thread.o 00:23:50.853 CXX test/cpp_headers/trace.o 00:23:51.110 CXX test/cpp_headers/trace_parser.o 00:23:51.110 CXX test/cpp_headers/tree.o 00:23:51.110 CXX test/cpp_headers/ublk.o 00:23:51.110 CXX test/cpp_headers/util.o 00:23:51.110 CXX test/cpp_headers/uuid.o 00:23:51.110 CXX test/cpp_headers/version.o 00:23:51.110 CXX test/cpp_headers/vfio_user_pci.o 00:23:51.110 CXX test/cpp_headers/vfio_user_spec.o 00:23:51.110 LINK nvmf 00:23:51.110 CXX test/cpp_headers/vhost.o 00:23:51.110 CXX test/cpp_headers/vmd.o 00:23:51.110 CXX test/cpp_headers/xor.o 00:23:51.110 CXX test/cpp_headers/zipf.o 00:23:52.043 LINK cuse 00:23:54.581 LINK esnap 00:23:54.581 ************************************ 00:23:54.581 END TEST make 00:23:54.581 ************************************ 00:23:54.581 00:23:54.581 real 1m12.196s 00:23:54.581 user 6m36.356s 00:23:54.581 sys 1m13.836s 00:23:54.581 13:44:01 make -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:23:54.581 13:44:01 make -- common/autotest_common.sh@10 -- $ set +x 00:23:54.581 13:44:01 -- spdk/autobuild.sh@1 -- $ stop_monitor_resources 00:23:54.581 13:44:01 -- pm/common@29 -- $ signal_monitor_resources TERM 00:23:54.581 13:44:01 -- pm/common@40 -- $ local monitor pid pids signal=TERM 00:23:54.581 13:44:01 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:23:54.581 13:44:01 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-cpu-load.pid ]] 00:23:54.581 13:44:01 -- pm/common@44 -- $ pid=5032 00:23:54.581 13:44:01 -- pm/common@50 -- $ kill -TERM 5032 00:23:54.581 13:44:01 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:23:54.581 13:44:01 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-vmstat.pid ]] 00:23:54.581 13:44:01 -- pm/common@44 -- $ pid=5033 00:23:54.581 13:44:01 -- pm/common@50 -- $ kill -TERM 5033 00:23:54.581 13:44:01 -- spdk/autorun.sh@26 -- $ (( SPDK_TEST_UNITTEST == 1 || SPDK_RUN_FUNCTIONAL_TEST == 1 )) 00:23:54.581 13:44:01 -- spdk/autorun.sh@27 -- $ sudo -E /home/vagrant/spdk_repo/spdk/autotest.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:23:54.581 13:44:01 -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:23:54.581 13:44:01 -- common/autotest_common.sh@1693 -- # lcov --version 00:23:54.581 13:44:01 -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:23:54.581 13:44:01 -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:23:54.581 13:44:01 -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:23:54.581 13:44:01 -- scripts/common.sh@333 -- # local ver1 ver1_l 00:23:54.581 13:44:01 -- scripts/common.sh@334 -- # local ver2 ver2_l 00:23:54.581 13:44:01 -- scripts/common.sh@336 -- # IFS=.-: 00:23:54.581 13:44:01 -- scripts/common.sh@336 -- # read -ra ver1 00:23:54.581 13:44:01 -- scripts/common.sh@337 -- # IFS=.-: 00:23:54.581 13:44:01 -- scripts/common.sh@337 -- # read -ra ver2 00:23:54.581 13:44:01 -- scripts/common.sh@338 -- # local 'op=<' 00:23:54.581 13:44:01 -- scripts/common.sh@340 -- # ver1_l=2 00:23:54.581 13:44:01 -- scripts/common.sh@341 -- # ver2_l=1 00:23:54.581 13:44:01 -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:23:54.581 13:44:01 -- scripts/common.sh@344 -- # case "$op" in 00:23:54.581 13:44:01 -- scripts/common.sh@345 -- # : 1 00:23:54.581 13:44:01 -- scripts/common.sh@364 -- # (( v = 0 )) 00:23:54.581 13:44:01 -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:23:54.581 13:44:01 -- scripts/common.sh@365 -- # decimal 1 00:23:54.581 13:44:01 -- scripts/common.sh@353 -- # local d=1 00:23:54.581 13:44:01 -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:23:54.581 13:44:01 -- scripts/common.sh@355 -- # echo 1 00:23:54.581 13:44:01 -- scripts/common.sh@365 -- # ver1[v]=1 00:23:54.581 13:44:01 -- scripts/common.sh@366 -- # decimal 2 00:23:54.581 13:44:01 -- scripts/common.sh@353 -- # local d=2 00:23:54.581 13:44:01 -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:23:54.581 13:44:01 -- scripts/common.sh@355 -- # echo 2 00:23:54.581 13:44:01 -- scripts/common.sh@366 -- # ver2[v]=2 00:23:54.581 13:44:01 -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:23:54.581 13:44:01 -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:23:54.581 13:44:01 -- scripts/common.sh@368 -- # return 0 00:23:54.581 13:44:01 -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:23:54.581 13:44:01 -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:23:54.581 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:54.581 --rc genhtml_branch_coverage=1 00:23:54.581 --rc genhtml_function_coverage=1 00:23:54.581 --rc genhtml_legend=1 00:23:54.581 --rc geninfo_all_blocks=1 00:23:54.581 --rc geninfo_unexecuted_blocks=1 00:23:54.581 00:23:54.581 ' 00:23:54.581 13:44:01 -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:23:54.581 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:54.581 --rc genhtml_branch_coverage=1 00:23:54.581 --rc genhtml_function_coverage=1 00:23:54.581 --rc genhtml_legend=1 00:23:54.581 --rc geninfo_all_blocks=1 00:23:54.581 --rc geninfo_unexecuted_blocks=1 00:23:54.581 00:23:54.581 ' 00:23:54.581 13:44:01 -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:23:54.581 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:54.581 --rc genhtml_branch_coverage=1 00:23:54.581 --rc genhtml_function_coverage=1 00:23:54.581 --rc genhtml_legend=1 00:23:54.581 --rc geninfo_all_blocks=1 00:23:54.581 --rc geninfo_unexecuted_blocks=1 00:23:54.581 00:23:54.581 ' 00:23:54.581 13:44:01 -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:23:54.581 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:23:54.581 --rc genhtml_branch_coverage=1 00:23:54.581 --rc genhtml_function_coverage=1 00:23:54.581 --rc genhtml_legend=1 00:23:54.581 --rc geninfo_all_blocks=1 00:23:54.581 --rc geninfo_unexecuted_blocks=1 00:23:54.581 00:23:54.581 ' 00:23:54.582 13:44:01 -- spdk/autotest.sh@25 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:23:54.582 13:44:01 -- nvmf/common.sh@7 -- # uname -s 00:23:54.582 13:44:01 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:23:54.582 13:44:01 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:23:54.582 13:44:01 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:23:54.582 13:44:01 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:23:54.582 13:44:01 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:23:54.582 13:44:01 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:23:54.582 13:44:01 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:23:54.582 13:44:01 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:23:54.582 13:44:01 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:23:54.582 13:44:01 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:23:54.582 13:44:01 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:0780082f-c229-467d-8bb4-b14a621a20b9 00:23:54.582 13:44:01 -- nvmf/common.sh@18 -- # NVME_HOSTID=0780082f-c229-467d-8bb4-b14a621a20b9 00:23:54.582 13:44:01 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:23:54.582 13:44:01 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:23:54.582 13:44:01 -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:23:54.582 13:44:01 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:23:54.582 13:44:01 -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:23:54.582 13:44:01 -- scripts/common.sh@15 -- # shopt -s extglob 00:23:54.582 13:44:01 -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:23:54.582 13:44:01 -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:23:54.582 13:44:01 -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:23:54.582 13:44:01 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:54.582 13:44:01 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:54.582 13:44:01 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:54.582 13:44:01 -- paths/export.sh@5 -- # export PATH 00:23:54.582 13:44:01 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:54.582 13:44:01 -- nvmf/common.sh@51 -- # : 0 00:23:54.582 13:44:01 -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:23:54.582 13:44:01 -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:23:54.582 13:44:01 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:23:54.582 13:44:01 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:23:54.582 13:44:01 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:23:54.582 13:44:01 -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:23:54.582 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:23:54.582 13:44:01 -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:23:54.582 13:44:01 -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:23:54.582 13:44:01 -- nvmf/common.sh@55 -- # have_pci_nics=0 00:23:54.582 13:44:01 -- spdk/autotest.sh@27 -- # '[' 0 -ne 0 ']' 00:23:54.582 13:44:01 -- spdk/autotest.sh@32 -- # uname -s 00:23:54.582 13:44:01 -- spdk/autotest.sh@32 -- # '[' Linux = Linux ']' 00:23:54.582 13:44:01 -- spdk/autotest.sh@33 -- # old_core_pattern='|/usr/lib/systemd/systemd-coredump %P %u %g %s %t %c %h' 00:23:54.582 13:44:01 -- spdk/autotest.sh@34 -- # mkdir -p /home/vagrant/spdk_repo/spdk/../output/coredumps 00:23:54.582 13:44:01 -- spdk/autotest.sh@39 -- # echo '|/home/vagrant/spdk_repo/spdk/scripts/core-collector.sh %P %s %t' 00:23:54.582 13:44:01 -- spdk/autotest.sh@40 -- # echo /home/vagrant/spdk_repo/spdk/../output/coredumps 00:23:54.582 13:44:01 -- spdk/autotest.sh@44 -- # modprobe nbd 00:23:54.840 13:44:01 -- spdk/autotest.sh@46 -- # type -P udevadm 00:23:54.840 13:44:01 -- spdk/autotest.sh@46 -- # udevadm=/usr/sbin/udevadm 00:23:54.840 13:44:01 -- spdk/autotest.sh@48 -- # udevadm_pid=53763 00:23:54.840 13:44:01 -- spdk/autotest.sh@53 -- # start_monitor_resources 00:23:54.840 13:44:01 -- pm/common@17 -- # local monitor 00:23:54.840 13:44:01 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:23:54.840 13:44:01 -- spdk/autotest.sh@47 -- # /usr/sbin/udevadm monitor --property 00:23:54.840 13:44:01 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:23:54.840 13:44:01 -- pm/common@25 -- # sleep 1 00:23:54.840 13:44:01 -- pm/common@21 -- # date +%s 00:23:54.840 13:44:01 -- pm/common@21 -- # date +%s 00:23:54.840 13:44:01 -- pm/common@21 -- # /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autotest.sh.1732110241 00:23:54.840 13:44:01 -- pm/common@21 -- # /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autotest.sh.1732110241 00:23:54.840 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autotest.sh.1732110241_collect-vmstat.pm.log 00:23:54.840 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autotest.sh.1732110241_collect-cpu-load.pm.log 00:23:55.790 13:44:02 -- spdk/autotest.sh@55 -- # trap 'autotest_cleanup || :; exit 1' SIGINT SIGTERM EXIT 00:23:55.790 13:44:02 -- spdk/autotest.sh@57 -- # timing_enter autotest 00:23:55.790 13:44:02 -- common/autotest_common.sh@726 -- # xtrace_disable 00:23:55.790 13:44:02 -- common/autotest_common.sh@10 -- # set +x 00:23:55.790 13:44:02 -- spdk/autotest.sh@59 -- # create_test_list 00:23:55.790 13:44:02 -- common/autotest_common.sh@752 -- # xtrace_disable 00:23:55.790 13:44:02 -- common/autotest_common.sh@10 -- # set +x 00:23:55.790 13:44:02 -- spdk/autotest.sh@61 -- # dirname /home/vagrant/spdk_repo/spdk/autotest.sh 00:23:55.790 13:44:02 -- spdk/autotest.sh@61 -- # readlink -f /home/vagrant/spdk_repo/spdk 00:23:55.790 13:44:02 -- spdk/autotest.sh@61 -- # src=/home/vagrant/spdk_repo/spdk 00:23:55.790 13:44:02 -- spdk/autotest.sh@62 -- # out=/home/vagrant/spdk_repo/spdk/../output 00:23:55.790 13:44:02 -- spdk/autotest.sh@63 -- # cd /home/vagrant/spdk_repo/spdk 00:23:55.790 13:44:02 -- spdk/autotest.sh@65 -- # freebsd_update_contigmem_mod 00:23:55.790 13:44:02 -- common/autotest_common.sh@1457 -- # uname 00:23:55.790 13:44:02 -- common/autotest_common.sh@1457 -- # '[' Linux = FreeBSD ']' 00:23:55.790 13:44:02 -- spdk/autotest.sh@66 -- # freebsd_set_maxsock_buf 00:23:55.790 13:44:02 -- common/autotest_common.sh@1477 -- # uname 00:23:55.790 13:44:02 -- common/autotest_common.sh@1477 -- # [[ Linux = FreeBSD ]] 00:23:55.790 13:44:02 -- spdk/autotest.sh@68 -- # [[ y == y ]] 00:23:55.790 13:44:02 -- spdk/autotest.sh@70 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 --version 00:23:55.790 lcov: LCOV version 1.15 00:23:55.790 13:44:02 -- spdk/autotest.sh@72 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -i -t Baseline -d /home/vagrant/spdk_repo/spdk -o /home/vagrant/spdk_repo/spdk/../output/cov_base.info 00:24:10.671 /home/vagrant/spdk_repo/spdk/lib/nvme/nvme_stubs.gcno:no functions found 00:24:10.671 geninfo: WARNING: GCOV did not produce any data for /home/vagrant/spdk_repo/spdk/lib/nvme/nvme_stubs.gcno 00:24:25.582 13:44:30 -- spdk/autotest.sh@76 -- # timing_enter pre_cleanup 00:24:25.582 13:44:30 -- common/autotest_common.sh@726 -- # xtrace_disable 00:24:25.582 13:44:30 -- common/autotest_common.sh@10 -- # set +x 00:24:25.582 13:44:30 -- spdk/autotest.sh@78 -- # rm -f 00:24:25.582 13:44:30 -- spdk/autotest.sh@81 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh reset 00:24:25.582 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:24:25.582 0000:00:11.0 (1b36 0010): Already using the nvme driver 00:24:25.582 0000:00:10.0 (1b36 0010): Already using the nvme driver 00:24:25.582 13:44:31 -- spdk/autotest.sh@83 -- # get_zoned_devs 00:24:25.582 13:44:31 -- common/autotest_common.sh@1657 -- # zoned_devs=() 00:24:25.582 13:44:31 -- common/autotest_common.sh@1657 -- # local -gA zoned_devs 00:24:25.582 13:44:31 -- common/autotest_common.sh@1658 -- # local nvme bdf 00:24:25.582 13:44:31 -- common/autotest_common.sh@1660 -- # for nvme in /sys/block/nvme* 00:24:25.582 13:44:31 -- common/autotest_common.sh@1661 -- # is_block_zoned nvme0n1 00:24:25.582 13:44:31 -- common/autotest_common.sh@1650 -- # local device=nvme0n1 00:24:25.582 13:44:31 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:24:25.582 13:44:31 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:24:25.582 13:44:31 -- common/autotest_common.sh@1660 -- # for nvme in /sys/block/nvme* 00:24:25.582 13:44:31 -- common/autotest_common.sh@1661 -- # is_block_zoned nvme1n1 00:24:25.582 13:44:31 -- common/autotest_common.sh@1650 -- # local device=nvme1n1 00:24:25.582 13:44:31 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme1n1/queue/zoned ]] 00:24:25.582 13:44:31 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:24:25.582 13:44:31 -- common/autotest_common.sh@1660 -- # for nvme in /sys/block/nvme* 00:24:25.582 13:44:31 -- common/autotest_common.sh@1661 -- # is_block_zoned nvme1n2 00:24:25.582 13:44:31 -- common/autotest_common.sh@1650 -- # local device=nvme1n2 00:24:25.582 13:44:31 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme1n2/queue/zoned ]] 00:24:25.582 13:44:31 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:24:25.582 13:44:31 -- common/autotest_common.sh@1660 -- # for nvme in /sys/block/nvme* 00:24:25.582 13:44:31 -- common/autotest_common.sh@1661 -- # is_block_zoned nvme1n3 00:24:25.582 13:44:31 -- common/autotest_common.sh@1650 -- # local device=nvme1n3 00:24:25.582 13:44:31 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme1n3/queue/zoned ]] 00:24:25.582 13:44:31 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:24:25.582 13:44:31 -- spdk/autotest.sh@85 -- # (( 0 > 0 )) 00:24:25.582 13:44:31 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:24:25.582 13:44:31 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:24:25.582 13:44:31 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme0n1 00:24:25.582 13:44:31 -- scripts/common.sh@381 -- # local block=/dev/nvme0n1 pt 00:24:25.582 13:44:31 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme0n1 00:24:25.582 No valid GPT data, bailing 00:24:25.582 13:44:31 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:24:25.582 13:44:31 -- scripts/common.sh@394 -- # pt= 00:24:25.582 13:44:31 -- scripts/common.sh@395 -- # return 1 00:24:25.582 13:44:31 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme0n1 bs=1M count=1 00:24:25.582 1+0 records in 00:24:25.582 1+0 records out 00:24:25.582 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00739863 s, 142 MB/s 00:24:25.582 13:44:31 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:24:25.582 13:44:31 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:24:25.582 13:44:31 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n1 00:24:25.582 13:44:31 -- scripts/common.sh@381 -- # local block=/dev/nvme1n1 pt 00:24:25.582 13:44:31 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n1 00:24:25.582 No valid GPT data, bailing 00:24:25.582 13:44:31 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n1 00:24:25.582 13:44:31 -- scripts/common.sh@394 -- # pt= 00:24:25.582 13:44:31 -- scripts/common.sh@395 -- # return 1 00:24:25.582 13:44:31 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n1 bs=1M count=1 00:24:25.582 1+0 records in 00:24:25.582 1+0 records out 00:24:25.582 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00514049 s, 204 MB/s 00:24:25.582 13:44:31 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:24:25.582 13:44:31 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:24:25.582 13:44:31 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n2 00:24:25.582 13:44:31 -- scripts/common.sh@381 -- # local block=/dev/nvme1n2 pt 00:24:25.582 13:44:31 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n2 00:24:25.582 No valid GPT data, bailing 00:24:25.582 13:44:31 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n2 00:24:25.582 13:44:31 -- scripts/common.sh@394 -- # pt= 00:24:25.582 13:44:31 -- scripts/common.sh@395 -- # return 1 00:24:25.582 13:44:31 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n2 bs=1M count=1 00:24:25.582 1+0 records in 00:24:25.582 1+0 records out 00:24:25.582 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00453237 s, 231 MB/s 00:24:25.582 13:44:31 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:24:25.582 13:44:31 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:24:25.582 13:44:31 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n3 00:24:25.582 13:44:31 -- scripts/common.sh@381 -- # local block=/dev/nvme1n3 pt 00:24:25.582 13:44:31 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n3 00:24:25.582 No valid GPT data, bailing 00:24:25.582 13:44:31 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n3 00:24:25.582 13:44:31 -- scripts/common.sh@394 -- # pt= 00:24:25.582 13:44:31 -- scripts/common.sh@395 -- # return 1 00:24:25.582 13:44:31 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n3 bs=1M count=1 00:24:25.582 1+0 records in 00:24:25.582 1+0 records out 00:24:25.582 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00303966 s, 345 MB/s 00:24:25.582 13:44:31 -- spdk/autotest.sh@105 -- # sync 00:24:25.583 13:44:31 -- spdk/autotest.sh@107 -- # xtrace_disable_per_cmd reap_spdk_processes 00:24:25.583 13:44:31 -- common/autotest_common.sh@22 -- # eval 'reap_spdk_processes 12> /dev/null' 00:24:25.583 13:44:31 -- common/autotest_common.sh@22 -- # reap_spdk_processes 00:24:26.153 13:44:33 -- spdk/autotest.sh@111 -- # uname -s 00:24:26.153 13:44:33 -- spdk/autotest.sh@111 -- # [[ Linux == Linux ]] 00:24:26.153 13:44:33 -- spdk/autotest.sh@111 -- # [[ 0 -eq 1 ]] 00:24:26.153 13:44:33 -- spdk/autotest.sh@115 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh status 00:24:26.719 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:24:26.719 Hugepages 00:24:26.719 node hugesize free / total 00:24:26.719 node0 1048576kB 0 / 0 00:24:26.719 node0 2048kB 0 / 0 00:24:26.719 00:24:26.719 Type BDF Vendor Device NUMA Driver Device Block devices 00:24:26.719 virtio 0000:00:03.0 1af4 1001 unknown virtio-pci - vda 00:24:26.976 NVMe 0000:00:10.0 1b36 0010 unknown nvme nvme0 nvme0n1 00:24:26.976 NVMe 0000:00:11.0 1b36 0010 unknown nvme nvme1 nvme1n1 nvme1n2 nvme1n3 00:24:26.977 13:44:33 -- spdk/autotest.sh@117 -- # uname -s 00:24:26.977 13:44:33 -- spdk/autotest.sh@117 -- # [[ Linux == Linux ]] 00:24:26.977 13:44:33 -- spdk/autotest.sh@119 -- # nvme_namespace_revert 00:24:26.977 13:44:33 -- common/autotest_common.sh@1516 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh 00:24:27.542 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:24:27.542 0000:00:10.0 (1b36 0010): nvme -> uio_pci_generic 00:24:27.542 0000:00:11.0 (1b36 0010): nvme -> uio_pci_generic 00:24:27.800 13:44:34 -- common/autotest_common.sh@1517 -- # sleep 1 00:24:28.732 13:44:35 -- common/autotest_common.sh@1518 -- # bdfs=() 00:24:28.732 13:44:35 -- common/autotest_common.sh@1518 -- # local bdfs 00:24:28.732 13:44:35 -- common/autotest_common.sh@1520 -- # bdfs=($(get_nvme_bdfs)) 00:24:28.732 13:44:35 -- common/autotest_common.sh@1520 -- # get_nvme_bdfs 00:24:28.732 13:44:35 -- common/autotest_common.sh@1498 -- # bdfs=() 00:24:28.732 13:44:35 -- common/autotest_common.sh@1498 -- # local bdfs 00:24:28.732 13:44:35 -- common/autotest_common.sh@1499 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:24:28.732 13:44:35 -- common/autotest_common.sh@1499 -- # jq -r '.config[].params.traddr' 00:24:28.732 13:44:35 -- common/autotest_common.sh@1499 -- # /home/vagrant/spdk_repo/spdk/scripts/gen_nvme.sh 00:24:28.732 13:44:35 -- common/autotest_common.sh@1500 -- # (( 2 == 0 )) 00:24:28.732 13:44:35 -- common/autotest_common.sh@1504 -- # printf '%s\n' 0000:00:10.0 0000:00:11.0 00:24:28.732 13:44:35 -- common/autotest_common.sh@1522 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh reset 00:24:28.991 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:24:28.991 Waiting for block devices as requested 00:24:28.991 0000:00:11.0 (1b36 0010): uio_pci_generic -> nvme 00:24:29.249 0000:00:10.0 (1b36 0010): uio_pci_generic -> nvme 00:24:29.249 13:44:36 -- common/autotest_common.sh@1524 -- # for bdf in "${bdfs[@]}" 00:24:29.249 13:44:36 -- common/autotest_common.sh@1525 -- # get_nvme_ctrlr_from_bdf 0000:00:10.0 00:24:29.249 13:44:36 -- common/autotest_common.sh@1487 -- # grep 0000:00:10.0/nvme/nvme 00:24:29.249 13:44:36 -- common/autotest_common.sh@1487 -- # readlink -f /sys/class/nvme/nvme0 /sys/class/nvme/nvme1 00:24:29.249 13:44:36 -- common/autotest_common.sh@1487 -- # bdf_sysfs_path=/sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 00:24:29.249 13:44:36 -- common/autotest_common.sh@1488 -- # [[ -z /sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 ]] 00:24:29.249 13:44:36 -- common/autotest_common.sh@1492 -- # basename /sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 00:24:29.249 13:44:36 -- common/autotest_common.sh@1492 -- # printf '%s\n' nvme1 00:24:29.249 13:44:36 -- common/autotest_common.sh@1525 -- # nvme_ctrlr=/dev/nvme1 00:24:29.249 13:44:36 -- common/autotest_common.sh@1526 -- # [[ -z /dev/nvme1 ]] 00:24:29.249 13:44:36 -- common/autotest_common.sh@1531 -- # nvme id-ctrl /dev/nvme1 00:24:29.249 13:44:36 -- common/autotest_common.sh@1531 -- # grep oacs 00:24:29.249 13:44:36 -- common/autotest_common.sh@1531 -- # cut -d: -f2 00:24:29.249 13:44:36 -- common/autotest_common.sh@1531 -- # oacs=' 0x12a' 00:24:29.249 13:44:36 -- common/autotest_common.sh@1532 -- # oacs_ns_manage=8 00:24:29.249 13:44:36 -- common/autotest_common.sh@1534 -- # [[ 8 -ne 0 ]] 00:24:29.249 13:44:36 -- common/autotest_common.sh@1540 -- # nvme id-ctrl /dev/nvme1 00:24:29.249 13:44:36 -- common/autotest_common.sh@1540 -- # grep unvmcap 00:24:29.249 13:44:36 -- common/autotest_common.sh@1540 -- # cut -d: -f2 00:24:29.249 13:44:36 -- common/autotest_common.sh@1540 -- # unvmcap=' 0' 00:24:29.249 13:44:36 -- common/autotest_common.sh@1541 -- # [[ 0 -eq 0 ]] 00:24:29.249 13:44:36 -- common/autotest_common.sh@1543 -- # continue 00:24:29.249 13:44:36 -- common/autotest_common.sh@1524 -- # for bdf in "${bdfs[@]}" 00:24:29.249 13:44:36 -- common/autotest_common.sh@1525 -- # get_nvme_ctrlr_from_bdf 0000:00:11.0 00:24:29.249 13:44:36 -- common/autotest_common.sh@1487 -- # readlink -f /sys/class/nvme/nvme0 /sys/class/nvme/nvme1 00:24:29.249 13:44:36 -- common/autotest_common.sh@1487 -- # grep 0000:00:11.0/nvme/nvme 00:24:29.249 13:44:36 -- common/autotest_common.sh@1487 -- # bdf_sysfs_path=/sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 00:24:29.249 13:44:36 -- common/autotest_common.sh@1488 -- # [[ -z /sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 ]] 00:24:29.249 13:44:36 -- common/autotest_common.sh@1492 -- # basename /sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 00:24:29.249 13:44:36 -- common/autotest_common.sh@1492 -- # printf '%s\n' nvme0 00:24:29.249 13:44:36 -- common/autotest_common.sh@1525 -- # nvme_ctrlr=/dev/nvme0 00:24:29.249 13:44:36 -- common/autotest_common.sh@1526 -- # [[ -z /dev/nvme0 ]] 00:24:29.249 13:44:36 -- common/autotest_common.sh@1531 -- # nvme id-ctrl /dev/nvme0 00:24:29.249 13:44:36 -- common/autotest_common.sh@1531 -- # grep oacs 00:24:29.249 13:44:36 -- common/autotest_common.sh@1531 -- # cut -d: -f2 00:24:29.249 13:44:36 -- common/autotest_common.sh@1531 -- # oacs=' 0x12a' 00:24:29.249 13:44:36 -- common/autotest_common.sh@1532 -- # oacs_ns_manage=8 00:24:29.249 13:44:36 -- common/autotest_common.sh@1534 -- # [[ 8 -ne 0 ]] 00:24:29.249 13:44:36 -- common/autotest_common.sh@1540 -- # nvme id-ctrl /dev/nvme0 00:24:29.249 13:44:36 -- common/autotest_common.sh@1540 -- # cut -d: -f2 00:24:29.249 13:44:36 -- common/autotest_common.sh@1540 -- # grep unvmcap 00:24:29.249 13:44:36 -- common/autotest_common.sh@1540 -- # unvmcap=' 0' 00:24:29.249 13:44:36 -- common/autotest_common.sh@1541 -- # [[ 0 -eq 0 ]] 00:24:29.249 13:44:36 -- common/autotest_common.sh@1543 -- # continue 00:24:29.250 13:44:36 -- spdk/autotest.sh@122 -- # timing_exit pre_cleanup 00:24:29.250 13:44:36 -- common/autotest_common.sh@732 -- # xtrace_disable 00:24:29.250 13:44:36 -- common/autotest_common.sh@10 -- # set +x 00:24:29.250 13:44:36 -- spdk/autotest.sh@125 -- # timing_enter afterboot 00:24:29.250 13:44:36 -- common/autotest_common.sh@726 -- # xtrace_disable 00:24:29.250 13:44:36 -- common/autotest_common.sh@10 -- # set +x 00:24:29.250 13:44:36 -- spdk/autotest.sh@126 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh 00:24:29.815 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:24:29.815 0000:00:10.0 (1b36 0010): nvme -> uio_pci_generic 00:24:30.073 0000:00:11.0 (1b36 0010): nvme -> uio_pci_generic 00:24:30.073 13:44:36 -- spdk/autotest.sh@127 -- # timing_exit afterboot 00:24:30.073 13:44:36 -- common/autotest_common.sh@732 -- # xtrace_disable 00:24:30.073 13:44:36 -- common/autotest_common.sh@10 -- # set +x 00:24:30.073 13:44:36 -- spdk/autotest.sh@131 -- # opal_revert_cleanup 00:24:30.073 13:44:36 -- common/autotest_common.sh@1578 -- # mapfile -t bdfs 00:24:30.073 13:44:36 -- common/autotest_common.sh@1578 -- # get_nvme_bdfs_by_id 0x0a54 00:24:30.073 13:44:36 -- common/autotest_common.sh@1563 -- # bdfs=() 00:24:30.073 13:44:36 -- common/autotest_common.sh@1563 -- # _bdfs=() 00:24:30.073 13:44:36 -- common/autotest_common.sh@1563 -- # local bdfs _bdfs 00:24:30.073 13:44:36 -- common/autotest_common.sh@1564 -- # _bdfs=($(get_nvme_bdfs)) 00:24:30.073 13:44:36 -- common/autotest_common.sh@1564 -- # get_nvme_bdfs 00:24:30.073 13:44:36 -- common/autotest_common.sh@1498 -- # bdfs=() 00:24:30.073 13:44:36 -- common/autotest_common.sh@1498 -- # local bdfs 00:24:30.073 13:44:36 -- common/autotest_common.sh@1499 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:24:30.073 13:44:36 -- common/autotest_common.sh@1499 -- # jq -r '.config[].params.traddr' 00:24:30.073 13:44:36 -- common/autotest_common.sh@1499 -- # /home/vagrant/spdk_repo/spdk/scripts/gen_nvme.sh 00:24:30.073 13:44:37 -- common/autotest_common.sh@1500 -- # (( 2 == 0 )) 00:24:30.073 13:44:37 -- common/autotest_common.sh@1504 -- # printf '%s\n' 0000:00:10.0 0000:00:11.0 00:24:30.073 13:44:37 -- common/autotest_common.sh@1565 -- # for bdf in "${_bdfs[@]}" 00:24:30.073 13:44:37 -- common/autotest_common.sh@1566 -- # cat /sys/bus/pci/devices/0000:00:10.0/device 00:24:30.073 13:44:37 -- common/autotest_common.sh@1566 -- # device=0x0010 00:24:30.073 13:44:37 -- common/autotest_common.sh@1567 -- # [[ 0x0010 == \0\x\0\a\5\4 ]] 00:24:30.073 13:44:37 -- common/autotest_common.sh@1565 -- # for bdf in "${_bdfs[@]}" 00:24:30.073 13:44:37 -- common/autotest_common.sh@1566 -- # cat /sys/bus/pci/devices/0000:00:11.0/device 00:24:30.073 13:44:37 -- common/autotest_common.sh@1566 -- # device=0x0010 00:24:30.073 13:44:37 -- common/autotest_common.sh@1567 -- # [[ 0x0010 == \0\x\0\a\5\4 ]] 00:24:30.073 13:44:37 -- common/autotest_common.sh@1572 -- # (( 0 > 0 )) 00:24:30.073 13:44:37 -- common/autotest_common.sh@1572 -- # return 0 00:24:30.073 13:44:37 -- common/autotest_common.sh@1579 -- # [[ -z '' ]] 00:24:30.073 13:44:37 -- common/autotest_common.sh@1580 -- # return 0 00:24:30.073 13:44:37 -- spdk/autotest.sh@137 -- # '[' 0 -eq 1 ']' 00:24:30.073 13:44:37 -- spdk/autotest.sh@141 -- # '[' 1 -eq 1 ']' 00:24:30.073 13:44:37 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:24:30.073 13:44:37 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:24:30.073 13:44:37 -- spdk/autotest.sh@149 -- # timing_enter lib 00:24:30.073 13:44:37 -- common/autotest_common.sh@726 -- # xtrace_disable 00:24:30.073 13:44:37 -- common/autotest_common.sh@10 -- # set +x 00:24:30.073 13:44:37 -- spdk/autotest.sh@151 -- # [[ 0 -eq 1 ]] 00:24:30.073 13:44:37 -- spdk/autotest.sh@155 -- # run_test env /home/vagrant/spdk_repo/spdk/test/env/env.sh 00:24:30.073 13:44:37 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:24:30.073 13:44:37 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:24:30.073 13:44:37 -- common/autotest_common.sh@10 -- # set +x 00:24:30.073 ************************************ 00:24:30.073 START TEST env 00:24:30.073 ************************************ 00:24:30.073 13:44:37 env -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/env.sh 00:24:30.073 * Looking for test storage... 00:24:30.073 * Found test storage at /home/vagrant/spdk_repo/spdk/test/env 00:24:30.073 13:44:37 env -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:24:30.073 13:44:37 env -- common/autotest_common.sh@1693 -- # lcov --version 00:24:30.073 13:44:37 env -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:24:30.331 13:44:37 env -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:24:30.331 13:44:37 env -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:24:30.331 13:44:37 env -- scripts/common.sh@333 -- # local ver1 ver1_l 00:24:30.331 13:44:37 env -- scripts/common.sh@334 -- # local ver2 ver2_l 00:24:30.331 13:44:37 env -- scripts/common.sh@336 -- # IFS=.-: 00:24:30.331 13:44:37 env -- scripts/common.sh@336 -- # read -ra ver1 00:24:30.331 13:44:37 env -- scripts/common.sh@337 -- # IFS=.-: 00:24:30.331 13:44:37 env -- scripts/common.sh@337 -- # read -ra ver2 00:24:30.331 13:44:37 env -- scripts/common.sh@338 -- # local 'op=<' 00:24:30.331 13:44:37 env -- scripts/common.sh@340 -- # ver1_l=2 00:24:30.331 13:44:37 env -- scripts/common.sh@341 -- # ver2_l=1 00:24:30.331 13:44:37 env -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:24:30.331 13:44:37 env -- scripts/common.sh@344 -- # case "$op" in 00:24:30.331 13:44:37 env -- scripts/common.sh@345 -- # : 1 00:24:30.331 13:44:37 env -- scripts/common.sh@364 -- # (( v = 0 )) 00:24:30.331 13:44:37 env -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:24:30.331 13:44:37 env -- scripts/common.sh@365 -- # decimal 1 00:24:30.331 13:44:37 env -- scripts/common.sh@353 -- # local d=1 00:24:30.331 13:44:37 env -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:24:30.331 13:44:37 env -- scripts/common.sh@355 -- # echo 1 00:24:30.331 13:44:37 env -- scripts/common.sh@365 -- # ver1[v]=1 00:24:30.331 13:44:37 env -- scripts/common.sh@366 -- # decimal 2 00:24:30.331 13:44:37 env -- scripts/common.sh@353 -- # local d=2 00:24:30.331 13:44:37 env -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:24:30.331 13:44:37 env -- scripts/common.sh@355 -- # echo 2 00:24:30.331 13:44:37 env -- scripts/common.sh@366 -- # ver2[v]=2 00:24:30.331 13:44:37 env -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:24:30.331 13:44:37 env -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:24:30.331 13:44:37 env -- scripts/common.sh@368 -- # return 0 00:24:30.331 13:44:37 env -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:24:30.331 13:44:37 env -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:24:30.331 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:30.331 --rc genhtml_branch_coverage=1 00:24:30.331 --rc genhtml_function_coverage=1 00:24:30.331 --rc genhtml_legend=1 00:24:30.331 --rc geninfo_all_blocks=1 00:24:30.331 --rc geninfo_unexecuted_blocks=1 00:24:30.331 00:24:30.331 ' 00:24:30.331 13:44:37 env -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:24:30.331 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:30.331 --rc genhtml_branch_coverage=1 00:24:30.331 --rc genhtml_function_coverage=1 00:24:30.331 --rc genhtml_legend=1 00:24:30.331 --rc geninfo_all_blocks=1 00:24:30.331 --rc geninfo_unexecuted_blocks=1 00:24:30.331 00:24:30.331 ' 00:24:30.331 13:44:37 env -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:24:30.331 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:30.331 --rc genhtml_branch_coverage=1 00:24:30.331 --rc genhtml_function_coverage=1 00:24:30.331 --rc genhtml_legend=1 00:24:30.331 --rc geninfo_all_blocks=1 00:24:30.331 --rc geninfo_unexecuted_blocks=1 00:24:30.331 00:24:30.331 ' 00:24:30.331 13:44:37 env -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:24:30.331 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:30.331 --rc genhtml_branch_coverage=1 00:24:30.331 --rc genhtml_function_coverage=1 00:24:30.331 --rc genhtml_legend=1 00:24:30.332 --rc geninfo_all_blocks=1 00:24:30.332 --rc geninfo_unexecuted_blocks=1 00:24:30.332 00:24:30.332 ' 00:24:30.332 13:44:37 env -- env/env.sh@10 -- # run_test env_memory /home/vagrant/spdk_repo/spdk/test/env/memory/memory_ut 00:24:30.332 13:44:37 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:24:30.332 13:44:37 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:24:30.332 13:44:37 env -- common/autotest_common.sh@10 -- # set +x 00:24:30.332 ************************************ 00:24:30.332 START TEST env_memory 00:24:30.332 ************************************ 00:24:30.332 13:44:37 env.env_memory -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/memory/memory_ut 00:24:30.332 00:24:30.332 00:24:30.332 CUnit - A unit testing framework for C - Version 2.1-3 00:24:30.332 http://cunit.sourceforge.net/ 00:24:30.332 00:24:30.332 00:24:30.332 Suite: memory 00:24:30.332 Test: alloc and free memory map ...[2024-11-20 13:44:37.259663] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 283:spdk_mem_map_alloc: *ERROR*: Initial mem_map notify failed 00:24:30.332 passed 00:24:30.332 Test: mem map translation ...[2024-11-20 13:44:37.299005] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=2097152 len=1234 00:24:30.332 [2024-11-20 13:44:37.299067] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=1234 len=2097152 00:24:30.332 [2024-11-20 13:44:37.299127] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 589:spdk_mem_map_set_translation: *ERROR*: invalid usermode virtual address 281474976710656 00:24:30.332 [2024-11-20 13:44:37.299144] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 605:spdk_mem_map_set_translation: *ERROR*: could not get 0xffffffe00000 map 00:24:30.332 passed 00:24:30.332 Test: mem map registration ...[2024-11-20 13:44:37.368156] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=200000 len=1234 00:24:30.332 [2024-11-20 13:44:37.368223] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=4d2 len=2097152 00:24:30.625 passed 00:24:30.625 Test: mem map adjacent registrations ...passed 00:24:30.625 00:24:30.625 Run Summary: Type Total Ran Passed Failed Inactive 00:24:30.625 suites 1 1 n/a 0 0 00:24:30.625 tests 4 4 4 0 0 00:24:30.625 asserts 152 152 152 0 n/a 00:24:30.625 00:24:30.625 Elapsed time = 0.293 seconds 00:24:30.625 00:24:30.625 real 0m0.328s 00:24:30.625 user 0m0.298s 00:24:30.625 sys 0m0.021s 00:24:30.625 13:44:37 env.env_memory -- common/autotest_common.sh@1130 -- # xtrace_disable 00:24:30.625 13:44:37 env.env_memory -- common/autotest_common.sh@10 -- # set +x 00:24:30.625 ************************************ 00:24:30.625 END TEST env_memory 00:24:30.625 ************************************ 00:24:30.625 13:44:37 env -- env/env.sh@11 -- # run_test env_vtophys /home/vagrant/spdk_repo/spdk/test/env/vtophys/vtophys 00:24:30.625 13:44:37 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:24:30.625 13:44:37 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:24:30.625 13:44:37 env -- common/autotest_common.sh@10 -- # set +x 00:24:30.625 ************************************ 00:24:30.625 START TEST env_vtophys 00:24:30.625 ************************************ 00:24:30.626 13:44:37 env.env_vtophys -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/vtophys/vtophys 00:24:30.626 EAL: lib.eal log level changed from notice to debug 00:24:30.626 EAL: Detected lcore 0 as core 0 on socket 0 00:24:30.626 EAL: Detected lcore 1 as core 0 on socket 0 00:24:30.626 EAL: Detected lcore 2 as core 0 on socket 0 00:24:30.626 EAL: Detected lcore 3 as core 0 on socket 0 00:24:30.626 EAL: Detected lcore 4 as core 0 on socket 0 00:24:30.626 EAL: Detected lcore 5 as core 0 on socket 0 00:24:30.626 EAL: Detected lcore 6 as core 0 on socket 0 00:24:30.626 EAL: Detected lcore 7 as core 0 on socket 0 00:24:30.626 EAL: Detected lcore 8 as core 0 on socket 0 00:24:30.626 EAL: Detected lcore 9 as core 0 on socket 0 00:24:30.626 EAL: Maximum logical cores by configuration: 128 00:24:30.626 EAL: Detected CPU lcores: 10 00:24:30.626 EAL: Detected NUMA nodes: 1 00:24:30.626 EAL: Checking presence of .so 'librte_eal.so.24.1' 00:24:30.626 EAL: Detected shared linkage of DPDK 00:24:30.626 EAL: No shared files mode enabled, IPC will be disabled 00:24:30.626 EAL: Selected IOVA mode 'PA' 00:24:30.626 EAL: Probing VFIO support... 00:24:30.626 EAL: Module /sys/module/vfio not found! error 2 (No such file or directory) 00:24:30.626 EAL: VFIO modules not loaded, skipping VFIO support... 00:24:30.626 EAL: Ask a virtual area of 0x2e000 bytes 00:24:30.626 EAL: Virtual area found at 0x200000000000 (size = 0x2e000) 00:24:30.626 EAL: Setting up physically contiguous memory... 00:24:30.626 EAL: Setting maximum number of open files to 524288 00:24:30.626 EAL: Detected memory type: socket_id:0 hugepage_sz:2097152 00:24:30.626 EAL: Creating 4 segment lists: n_segs:8192 socket_id:0 hugepage_sz:2097152 00:24:30.626 EAL: Ask a virtual area of 0x61000 bytes 00:24:30.626 EAL: Virtual area found at 0x20000002e000 (size = 0x61000) 00:24:30.626 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:24:30.626 EAL: Ask a virtual area of 0x400000000 bytes 00:24:30.626 EAL: Virtual area found at 0x200000200000 (size = 0x400000000) 00:24:30.626 EAL: VA reserved for memseg list at 0x200000200000, size 400000000 00:24:30.626 EAL: Ask a virtual area of 0x61000 bytes 00:24:30.626 EAL: Virtual area found at 0x200400200000 (size = 0x61000) 00:24:30.626 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:24:30.626 EAL: Ask a virtual area of 0x400000000 bytes 00:24:30.626 EAL: Virtual area found at 0x200400400000 (size = 0x400000000) 00:24:30.626 EAL: VA reserved for memseg list at 0x200400400000, size 400000000 00:24:30.626 EAL: Ask a virtual area of 0x61000 bytes 00:24:30.626 EAL: Virtual area found at 0x200800400000 (size = 0x61000) 00:24:30.626 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:24:30.626 EAL: Ask a virtual area of 0x400000000 bytes 00:24:30.626 EAL: Virtual area found at 0x200800600000 (size = 0x400000000) 00:24:30.626 EAL: VA reserved for memseg list at 0x200800600000, size 400000000 00:24:30.626 EAL: Ask a virtual area of 0x61000 bytes 00:24:30.626 EAL: Virtual area found at 0x200c00600000 (size = 0x61000) 00:24:30.626 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:24:30.626 EAL: Ask a virtual area of 0x400000000 bytes 00:24:30.626 EAL: Virtual area found at 0x200c00800000 (size = 0x400000000) 00:24:30.626 EAL: VA reserved for memseg list at 0x200c00800000, size 400000000 00:24:30.626 EAL: Hugepages will be freed exactly as allocated. 00:24:30.626 EAL: No shared files mode enabled, IPC is disabled 00:24:30.626 EAL: No shared files mode enabled, IPC is disabled 00:24:30.899 EAL: TSC frequency is ~2600000 KHz 00:24:30.899 EAL: Main lcore 0 is ready (tid=7f22c9aefa40;cpuset=[0]) 00:24:30.899 EAL: Trying to obtain current memory policy. 00:24:30.899 EAL: Setting policy MPOL_PREFERRED for socket 0 00:24:30.899 EAL: Restoring previous memory policy: 0 00:24:30.899 EAL: request: mp_malloc_sync 00:24:30.899 EAL: No shared files mode enabled, IPC is disabled 00:24:30.899 EAL: Heap on socket 0 was expanded by 2MB 00:24:30.899 EAL: Module /sys/module/vfio not found! error 2 (No such file or directory) 00:24:30.899 EAL: No PCI address specified using 'addr=' in: bus=pci 00:24:30.899 EAL: Mem event callback 'spdk:(nil)' registered 00:24:30.899 EAL: Module /sys/module/vfio_pci not found! error 2 (No such file or directory) 00:24:30.899 00:24:30.899 00:24:30.899 CUnit - A unit testing framework for C - Version 2.1-3 00:24:30.899 http://cunit.sourceforge.net/ 00:24:30.899 00:24:30.899 00:24:30.899 Suite: components_suite 00:24:31.159 Test: vtophys_malloc_test ...passed 00:24:31.159 Test: vtophys_spdk_malloc_test ...EAL: Trying to obtain current memory policy. 00:24:31.159 EAL: Setting policy MPOL_PREFERRED for socket 0 00:24:31.159 EAL: Restoring previous memory policy: 4 00:24:31.159 EAL: Calling mem event callback 'spdk:(nil)' 00:24:31.159 EAL: request: mp_malloc_sync 00:24:31.159 EAL: No shared files mode enabled, IPC is disabled 00:24:31.159 EAL: Heap on socket 0 was expanded by 4MB 00:24:31.159 EAL: Calling mem event callback 'spdk:(nil)' 00:24:31.159 EAL: request: mp_malloc_sync 00:24:31.159 EAL: No shared files mode enabled, IPC is disabled 00:24:31.159 EAL: Heap on socket 0 was shrunk by 4MB 00:24:31.159 EAL: Trying to obtain current memory policy. 00:24:31.159 EAL: Setting policy MPOL_PREFERRED for socket 0 00:24:31.159 EAL: Restoring previous memory policy: 4 00:24:31.159 EAL: Calling mem event callback 'spdk:(nil)' 00:24:31.159 EAL: request: mp_malloc_sync 00:24:31.159 EAL: No shared files mode enabled, IPC is disabled 00:24:31.159 EAL: Heap on socket 0 was expanded by 6MB 00:24:31.159 EAL: Calling mem event callback 'spdk:(nil)' 00:24:31.159 EAL: request: mp_malloc_sync 00:24:31.159 EAL: No shared files mode enabled, IPC is disabled 00:24:31.159 EAL: Heap on socket 0 was shrunk by 6MB 00:24:31.159 EAL: Trying to obtain current memory policy. 00:24:31.159 EAL: Setting policy MPOL_PREFERRED for socket 0 00:24:31.159 EAL: Restoring previous memory policy: 4 00:24:31.159 EAL: Calling mem event callback 'spdk:(nil)' 00:24:31.159 EAL: request: mp_malloc_sync 00:24:31.159 EAL: No shared files mode enabled, IPC is disabled 00:24:31.159 EAL: Heap on socket 0 was expanded by 10MB 00:24:31.159 EAL: Calling mem event callback 'spdk:(nil)' 00:24:31.159 EAL: request: mp_malloc_sync 00:24:31.159 EAL: No shared files mode enabled, IPC is disabled 00:24:31.159 EAL: Heap on socket 0 was shrunk by 10MB 00:24:31.159 EAL: Trying to obtain current memory policy. 00:24:31.159 EAL: Setting policy MPOL_PREFERRED for socket 0 00:24:31.159 EAL: Restoring previous memory policy: 4 00:24:31.159 EAL: Calling mem event callback 'spdk:(nil)' 00:24:31.159 EAL: request: mp_malloc_sync 00:24:31.159 EAL: No shared files mode enabled, IPC is disabled 00:24:31.159 EAL: Heap on socket 0 was expanded by 18MB 00:24:31.159 EAL: Calling mem event callback 'spdk:(nil)' 00:24:31.159 EAL: request: mp_malloc_sync 00:24:31.159 EAL: No shared files mode enabled, IPC is disabled 00:24:31.159 EAL: Heap on socket 0 was shrunk by 18MB 00:24:31.159 EAL: Trying to obtain current memory policy. 00:24:31.159 EAL: Setting policy MPOL_PREFERRED for socket 0 00:24:31.159 EAL: Restoring previous memory policy: 4 00:24:31.159 EAL: Calling mem event callback 'spdk:(nil)' 00:24:31.159 EAL: request: mp_malloc_sync 00:24:31.159 EAL: No shared files mode enabled, IPC is disabled 00:24:31.159 EAL: Heap on socket 0 was expanded by 34MB 00:24:31.418 EAL: Calling mem event callback 'spdk:(nil)' 00:24:31.418 EAL: request: mp_malloc_sync 00:24:31.418 EAL: No shared files mode enabled, IPC is disabled 00:24:31.418 EAL: Heap on socket 0 was shrunk by 34MB 00:24:31.418 EAL: Trying to obtain current memory policy. 00:24:31.418 EAL: Setting policy MPOL_PREFERRED for socket 0 00:24:31.418 EAL: Restoring previous memory policy: 4 00:24:31.418 EAL: Calling mem event callback 'spdk:(nil)' 00:24:31.418 EAL: request: mp_malloc_sync 00:24:31.418 EAL: No shared files mode enabled, IPC is disabled 00:24:31.418 EAL: Heap on socket 0 was expanded by 66MB 00:24:31.418 EAL: Calling mem event callback 'spdk:(nil)' 00:24:31.418 EAL: request: mp_malloc_sync 00:24:31.418 EAL: No shared files mode enabled, IPC is disabled 00:24:31.418 EAL: Heap on socket 0 was shrunk by 66MB 00:24:31.418 EAL: Trying to obtain current memory policy. 00:24:31.418 EAL: Setting policy MPOL_PREFERRED for socket 0 00:24:31.418 EAL: Restoring previous memory policy: 4 00:24:31.418 EAL: Calling mem event callback 'spdk:(nil)' 00:24:31.418 EAL: request: mp_malloc_sync 00:24:31.418 EAL: No shared files mode enabled, IPC is disabled 00:24:31.418 EAL: Heap on socket 0 was expanded by 130MB 00:24:31.676 EAL: Calling mem event callback 'spdk:(nil)' 00:24:31.676 EAL: request: mp_malloc_sync 00:24:31.676 EAL: No shared files mode enabled, IPC is disabled 00:24:31.676 EAL: Heap on socket 0 was shrunk by 130MB 00:24:31.676 EAL: Trying to obtain current memory policy. 00:24:31.676 EAL: Setting policy MPOL_PREFERRED for socket 0 00:24:31.934 EAL: Restoring previous memory policy: 4 00:24:31.934 EAL: Calling mem event callback 'spdk:(nil)' 00:24:31.934 EAL: request: mp_malloc_sync 00:24:31.934 EAL: No shared files mode enabled, IPC is disabled 00:24:31.934 EAL: Heap on socket 0 was expanded by 258MB 00:24:32.193 EAL: Calling mem event callback 'spdk:(nil)' 00:24:32.193 EAL: request: mp_malloc_sync 00:24:32.193 EAL: No shared files mode enabled, IPC is disabled 00:24:32.193 EAL: Heap on socket 0 was shrunk by 258MB 00:24:32.451 EAL: Trying to obtain current memory policy. 00:24:32.451 EAL: Setting policy MPOL_PREFERRED for socket 0 00:24:32.451 EAL: Restoring previous memory policy: 4 00:24:32.451 EAL: Calling mem event callback 'spdk:(nil)' 00:24:32.451 EAL: request: mp_malloc_sync 00:24:32.451 EAL: No shared files mode enabled, IPC is disabled 00:24:32.451 EAL: Heap on socket 0 was expanded by 514MB 00:24:33.017 EAL: Calling mem event callback 'spdk:(nil)' 00:24:33.017 EAL: request: mp_malloc_sync 00:24:33.017 EAL: No shared files mode enabled, IPC is disabled 00:24:33.017 EAL: Heap on socket 0 was shrunk by 514MB 00:24:33.583 EAL: Trying to obtain current memory policy. 00:24:33.583 EAL: Setting policy MPOL_PREFERRED for socket 0 00:24:33.842 EAL: Restoring previous memory policy: 4 00:24:33.842 EAL: Calling mem event callback 'spdk:(nil)' 00:24:33.842 EAL: request: mp_malloc_sync 00:24:33.842 EAL: No shared files mode enabled, IPC is disabled 00:24:33.842 EAL: Heap on socket 0 was expanded by 1026MB 00:24:34.777 EAL: Calling mem event callback 'spdk:(nil)' 00:24:34.777 EAL: request: mp_malloc_sync 00:24:34.777 EAL: No shared files mode enabled, IPC is disabled 00:24:34.777 EAL: Heap on socket 0 was shrunk by 1026MB 00:24:35.710 passed 00:24:35.710 00:24:35.710 Run Summary: Type Total Ran Passed Failed Inactive 00:24:35.711 suites 1 1 n/a 0 0 00:24:35.711 tests 2 2 2 0 0 00:24:35.711 asserts 5831 5831 5831 0 n/a 00:24:35.711 00:24:35.711 Elapsed time = 4.708 seconds 00:24:35.711 EAL: Calling mem event callback 'spdk:(nil)' 00:24:35.711 EAL: request: mp_malloc_sync 00:24:35.711 EAL: No shared files mode enabled, IPC is disabled 00:24:35.711 EAL: Heap on socket 0 was shrunk by 2MB 00:24:35.711 EAL: No shared files mode enabled, IPC is disabled 00:24:35.711 EAL: No shared files mode enabled, IPC is disabled 00:24:35.711 EAL: No shared files mode enabled, IPC is disabled 00:24:35.711 00:24:35.711 real 0m4.977s 00:24:35.711 user 0m4.113s 00:24:35.711 sys 0m0.716s 00:24:35.711 13:44:42 env.env_vtophys -- common/autotest_common.sh@1130 -- # xtrace_disable 00:24:35.711 13:44:42 env.env_vtophys -- common/autotest_common.sh@10 -- # set +x 00:24:35.711 ************************************ 00:24:35.711 END TEST env_vtophys 00:24:35.711 ************************************ 00:24:35.711 13:44:42 env -- env/env.sh@12 -- # run_test env_pci /home/vagrant/spdk_repo/spdk/test/env/pci/pci_ut 00:24:35.711 13:44:42 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:24:35.711 13:44:42 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:24:35.711 13:44:42 env -- common/autotest_common.sh@10 -- # set +x 00:24:35.711 ************************************ 00:24:35.711 START TEST env_pci 00:24:35.711 ************************************ 00:24:35.711 13:44:42 env.env_pci -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/pci/pci_ut 00:24:35.711 00:24:35.711 00:24:35.711 CUnit - A unit testing framework for C - Version 2.1-3 00:24:35.711 http://cunit.sourceforge.net/ 00:24:35.711 00:24:35.711 00:24:35.711 Suite: pci 00:24:35.711 Test: pci_hook ...[2024-11-20 13:44:42.615849] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/pci.c:1117:spdk_pci_device_claim: *ERROR*: Cannot create lock on device /var/tmp/spdk_pci_lock_10000:00:01.0, probably process 55982 has claimed it 00:24:35.711 passedEAL: Cannot find device (10000:00:01.0) 00:24:35.711 EAL: Failed to attach device on primary process 00:24:35.711 00:24:35.711 00:24:35.711 Run Summary: Type Total Ran Passed Failed Inactive 00:24:35.711 suites 1 1 n/a 0 0 00:24:35.711 tests 1 1 1 0 0 00:24:35.711 asserts 25 25 25 0 n/a 00:24:35.711 00:24:35.711 Elapsed time = 0.006 seconds 00:24:35.711 00:24:35.711 real 0m0.058s 00:24:35.711 user 0m0.021s 00:24:35.711 sys 0m0.037s 00:24:35.711 13:44:42 env.env_pci -- common/autotest_common.sh@1130 -- # xtrace_disable 00:24:35.711 13:44:42 env.env_pci -- common/autotest_common.sh@10 -- # set +x 00:24:35.711 ************************************ 00:24:35.711 END TEST env_pci 00:24:35.711 ************************************ 00:24:35.711 13:44:42 env -- env/env.sh@14 -- # argv='-c 0x1 ' 00:24:35.711 13:44:42 env -- env/env.sh@15 -- # uname 00:24:35.711 13:44:42 env -- env/env.sh@15 -- # '[' Linux = Linux ']' 00:24:35.711 13:44:42 env -- env/env.sh@22 -- # argv+=--base-virtaddr=0x200000000000 00:24:35.711 13:44:42 env -- env/env.sh@24 -- # run_test env_dpdk_post_init /home/vagrant/spdk_repo/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:24:35.711 13:44:42 env -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:24:35.711 13:44:42 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:24:35.711 13:44:42 env -- common/autotest_common.sh@10 -- # set +x 00:24:35.711 ************************************ 00:24:35.711 START TEST env_dpdk_post_init 00:24:35.711 ************************************ 00:24:35.711 13:44:42 env.env_dpdk_post_init -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:24:35.711 EAL: Detected CPU lcores: 10 00:24:35.711 EAL: Detected NUMA nodes: 1 00:24:35.711 EAL: Detected shared linkage of DPDK 00:24:35.711 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:24:35.711 EAL: Selected IOVA mode 'PA' 00:24:35.969 TELEMETRY: No legacy callbacks, legacy socket not created 00:24:35.969 EAL: Probe PCI driver: spdk_nvme (1b36:0010) device: 0000:00:10.0 (socket -1) 00:24:35.969 EAL: Probe PCI driver: spdk_nvme (1b36:0010) device: 0000:00:11.0 (socket -1) 00:24:35.969 Starting DPDK initialization... 00:24:35.969 Starting SPDK post initialization... 00:24:35.969 SPDK NVMe probe 00:24:35.969 Attaching to 0000:00:10.0 00:24:35.969 Attaching to 0000:00:11.0 00:24:35.969 Attached to 0000:00:10.0 00:24:35.969 Attached to 0000:00:11.0 00:24:35.969 Cleaning up... 00:24:35.969 00:24:35.969 real 0m0.238s 00:24:35.969 user 0m0.078s 00:24:35.969 sys 0m0.061s 00:24:35.969 13:44:42 env.env_dpdk_post_init -- common/autotest_common.sh@1130 -- # xtrace_disable 00:24:35.969 13:44:42 env.env_dpdk_post_init -- common/autotest_common.sh@10 -- # set +x 00:24:35.969 ************************************ 00:24:35.969 END TEST env_dpdk_post_init 00:24:35.969 ************************************ 00:24:35.969 13:44:42 env -- env/env.sh@26 -- # uname 00:24:35.969 13:44:42 env -- env/env.sh@26 -- # '[' Linux = Linux ']' 00:24:35.969 13:44:42 env -- env/env.sh@29 -- # run_test env_mem_callbacks /home/vagrant/spdk_repo/spdk/test/env/mem_callbacks/mem_callbacks 00:24:35.969 13:44:42 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:24:35.969 13:44:42 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:24:35.969 13:44:42 env -- common/autotest_common.sh@10 -- # set +x 00:24:35.969 ************************************ 00:24:35.969 START TEST env_mem_callbacks 00:24:35.969 ************************************ 00:24:35.969 13:44:42 env.env_mem_callbacks -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/mem_callbacks/mem_callbacks 00:24:35.969 EAL: Detected CPU lcores: 10 00:24:35.969 EAL: Detected NUMA nodes: 1 00:24:35.969 EAL: Detected shared linkage of DPDK 00:24:35.969 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:24:36.226 EAL: Selected IOVA mode 'PA' 00:24:36.226 TELEMETRY: No legacy callbacks, legacy socket not created 00:24:36.226 00:24:36.226 00:24:36.226 CUnit - A unit testing framework for C - Version 2.1-3 00:24:36.226 http://cunit.sourceforge.net/ 00:24:36.226 00:24:36.226 00:24:36.226 Suite: memory 00:24:36.226 Test: test ... 00:24:36.226 register 0x200000200000 2097152 00:24:36.226 malloc 3145728 00:24:36.226 register 0x200000400000 4194304 00:24:36.226 buf 0x2000004fffc0 len 3145728 PASSED 00:24:36.226 malloc 64 00:24:36.226 buf 0x2000004ffec0 len 64 PASSED 00:24:36.226 malloc 4194304 00:24:36.226 register 0x200000800000 6291456 00:24:36.226 buf 0x2000009fffc0 len 4194304 PASSED 00:24:36.226 free 0x2000004fffc0 3145728 00:24:36.226 free 0x2000004ffec0 64 00:24:36.226 unregister 0x200000400000 4194304 PASSED 00:24:36.226 free 0x2000009fffc0 4194304 00:24:36.226 unregister 0x200000800000 6291456 PASSED 00:24:36.226 malloc 8388608 00:24:36.226 register 0x200000400000 10485760 00:24:36.226 buf 0x2000005fffc0 len 8388608 PASSED 00:24:36.226 free 0x2000005fffc0 8388608 00:24:36.226 unregister 0x200000400000 10485760 PASSED 00:24:36.226 passed 00:24:36.226 00:24:36.226 Run Summary: Type Total Ran Passed Failed Inactive 00:24:36.226 suites 1 1 n/a 0 0 00:24:36.226 tests 1 1 1 0 0 00:24:36.226 asserts 15 15 15 0 n/a 00:24:36.226 00:24:36.226 Elapsed time = 0.052 seconds 00:24:36.226 00:24:36.226 real 0m0.217s 00:24:36.226 user 0m0.069s 00:24:36.226 sys 0m0.047s 00:24:36.226 13:44:43 env.env_mem_callbacks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:24:36.226 13:44:43 env.env_mem_callbacks -- common/autotest_common.sh@10 -- # set +x 00:24:36.226 ************************************ 00:24:36.226 END TEST env_mem_callbacks 00:24:36.226 ************************************ 00:24:36.226 00:24:36.226 real 0m6.181s 00:24:36.226 user 0m4.734s 00:24:36.226 sys 0m1.091s 00:24:36.226 13:44:43 env -- common/autotest_common.sh@1130 -- # xtrace_disable 00:24:36.226 13:44:43 env -- common/autotest_common.sh@10 -- # set +x 00:24:36.226 ************************************ 00:24:36.226 END TEST env 00:24:36.226 ************************************ 00:24:36.226 13:44:43 -- spdk/autotest.sh@156 -- # run_test rpc /home/vagrant/spdk_repo/spdk/test/rpc/rpc.sh 00:24:36.226 13:44:43 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:24:36.226 13:44:43 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:24:36.226 13:44:43 -- common/autotest_common.sh@10 -- # set +x 00:24:36.226 ************************************ 00:24:36.226 START TEST rpc 00:24:36.226 ************************************ 00:24:36.226 13:44:43 rpc -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/rpc/rpc.sh 00:24:36.483 * Looking for test storage... 00:24:36.483 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc 00:24:36.483 13:44:43 rpc -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:24:36.483 13:44:43 rpc -- common/autotest_common.sh@1693 -- # lcov --version 00:24:36.483 13:44:43 rpc -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:24:36.483 13:44:43 rpc -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:24:36.483 13:44:43 rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:24:36.484 13:44:43 rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:24:36.484 13:44:43 rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:24:36.484 13:44:43 rpc -- scripts/common.sh@336 -- # IFS=.-: 00:24:36.484 13:44:43 rpc -- scripts/common.sh@336 -- # read -ra ver1 00:24:36.484 13:44:43 rpc -- scripts/common.sh@337 -- # IFS=.-: 00:24:36.484 13:44:43 rpc -- scripts/common.sh@337 -- # read -ra ver2 00:24:36.484 13:44:43 rpc -- scripts/common.sh@338 -- # local 'op=<' 00:24:36.484 13:44:43 rpc -- scripts/common.sh@340 -- # ver1_l=2 00:24:36.484 13:44:43 rpc -- scripts/common.sh@341 -- # ver2_l=1 00:24:36.484 13:44:43 rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:24:36.484 13:44:43 rpc -- scripts/common.sh@344 -- # case "$op" in 00:24:36.484 13:44:43 rpc -- scripts/common.sh@345 -- # : 1 00:24:36.484 13:44:43 rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:24:36.484 13:44:43 rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:24:36.484 13:44:43 rpc -- scripts/common.sh@365 -- # decimal 1 00:24:36.484 13:44:43 rpc -- scripts/common.sh@353 -- # local d=1 00:24:36.484 13:44:43 rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:24:36.484 13:44:43 rpc -- scripts/common.sh@355 -- # echo 1 00:24:36.484 13:44:43 rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:24:36.484 13:44:43 rpc -- scripts/common.sh@366 -- # decimal 2 00:24:36.484 13:44:43 rpc -- scripts/common.sh@353 -- # local d=2 00:24:36.484 13:44:43 rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:24:36.484 13:44:43 rpc -- scripts/common.sh@355 -- # echo 2 00:24:36.484 13:44:43 rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:24:36.484 13:44:43 rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:24:36.484 13:44:43 rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:24:36.484 13:44:43 rpc -- scripts/common.sh@368 -- # return 0 00:24:36.484 13:44:43 rpc -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:24:36.484 13:44:43 rpc -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:24:36.484 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:36.484 --rc genhtml_branch_coverage=1 00:24:36.484 --rc genhtml_function_coverage=1 00:24:36.484 --rc genhtml_legend=1 00:24:36.484 --rc geninfo_all_blocks=1 00:24:36.484 --rc geninfo_unexecuted_blocks=1 00:24:36.484 00:24:36.484 ' 00:24:36.484 13:44:43 rpc -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:24:36.484 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:36.484 --rc genhtml_branch_coverage=1 00:24:36.484 --rc genhtml_function_coverage=1 00:24:36.484 --rc genhtml_legend=1 00:24:36.484 --rc geninfo_all_blocks=1 00:24:36.484 --rc geninfo_unexecuted_blocks=1 00:24:36.484 00:24:36.484 ' 00:24:36.484 13:44:43 rpc -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:24:36.484 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:36.484 --rc genhtml_branch_coverage=1 00:24:36.484 --rc genhtml_function_coverage=1 00:24:36.484 --rc genhtml_legend=1 00:24:36.484 --rc geninfo_all_blocks=1 00:24:36.484 --rc geninfo_unexecuted_blocks=1 00:24:36.484 00:24:36.484 ' 00:24:36.484 13:44:43 rpc -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:24:36.484 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:36.484 --rc genhtml_branch_coverage=1 00:24:36.484 --rc genhtml_function_coverage=1 00:24:36.484 --rc genhtml_legend=1 00:24:36.484 --rc geninfo_all_blocks=1 00:24:36.484 --rc geninfo_unexecuted_blocks=1 00:24:36.484 00:24:36.484 ' 00:24:36.484 13:44:43 rpc -- rpc/rpc.sh@65 -- # spdk_pid=56103 00:24:36.484 13:44:43 rpc -- rpc/rpc.sh@66 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:24:36.484 13:44:43 rpc -- rpc/rpc.sh@64 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -e bdev 00:24:36.484 13:44:43 rpc -- rpc/rpc.sh@67 -- # waitforlisten 56103 00:24:36.484 13:44:43 rpc -- common/autotest_common.sh@835 -- # '[' -z 56103 ']' 00:24:36.484 13:44:43 rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:36.484 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:36.484 13:44:43 rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:24:36.484 13:44:43 rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:36.484 13:44:43 rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:24:36.484 13:44:43 rpc -- common/autotest_common.sh@10 -- # set +x 00:24:36.484 [2024-11-20 13:44:43.530732] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:24:36.484 [2024-11-20 13:44:43.530912] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid56103 ] 00:24:36.741 [2024-11-20 13:44:43.691045] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:36.998 [2024-11-20 13:44:43.801419] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask bdev specified. 00:24:36.998 [2024-11-20 13:44:43.801547] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s spdk_tgt -p 56103' to capture a snapshot of events at runtime. 00:24:36.998 [2024-11-20 13:44:43.801577] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:24:36.998 [2024-11-20 13:44:43.801609] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:24:36.998 [2024-11-20 13:44:43.801631] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/spdk_tgt_trace.pid56103 for offline analysis/debug. 00:24:36.998 [2024-11-20 13:44:43.803762] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:24:37.562 13:44:44 rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:24:37.562 13:44:44 rpc -- common/autotest_common.sh@868 -- # return 0 00:24:37.562 13:44:44 rpc -- rpc/rpc.sh@69 -- # export PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/test/rpc 00:24:37.562 13:44:44 rpc -- rpc/rpc.sh@69 -- # PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/test/rpc 00:24:37.562 13:44:44 rpc -- rpc/rpc.sh@72 -- # rpc=rpc_cmd 00:24:37.562 13:44:44 rpc -- rpc/rpc.sh@73 -- # run_test rpc_integrity rpc_integrity 00:24:37.562 13:44:44 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:24:37.562 13:44:44 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:24:37.562 13:44:44 rpc -- common/autotest_common.sh@10 -- # set +x 00:24:37.562 ************************************ 00:24:37.562 START TEST rpc_integrity 00:24:37.562 ************************************ 00:24:37.562 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@1129 -- # rpc_integrity 00:24:37.562 13:44:44 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:24:37.562 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:37.562 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:24:37.562 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:37.562 13:44:44 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:24:37.562 13:44:44 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # jq length 00:24:37.562 13:44:44 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:24:37.562 13:44:44 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:24:37.562 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:37.562 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:24:37.562 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:37.562 13:44:44 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc0 00:24:37.562 13:44:44 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:24:37.562 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:37.562 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:24:37.562 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:37.562 13:44:44 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:24:37.562 { 00:24:37.562 "name": "Malloc0", 00:24:37.562 "aliases": [ 00:24:37.562 "97a267fa-718c-4c9a-a372-614170c633f1" 00:24:37.562 ], 00:24:37.562 "product_name": "Malloc disk", 00:24:37.562 "block_size": 512, 00:24:37.562 "num_blocks": 16384, 00:24:37.562 "uuid": "97a267fa-718c-4c9a-a372-614170c633f1", 00:24:37.562 "assigned_rate_limits": { 00:24:37.562 "rw_ios_per_sec": 0, 00:24:37.562 "rw_mbytes_per_sec": 0, 00:24:37.562 "r_mbytes_per_sec": 0, 00:24:37.562 "w_mbytes_per_sec": 0 00:24:37.562 }, 00:24:37.562 "claimed": false, 00:24:37.562 "zoned": false, 00:24:37.562 "supported_io_types": { 00:24:37.562 "read": true, 00:24:37.562 "write": true, 00:24:37.562 "unmap": true, 00:24:37.562 "flush": true, 00:24:37.562 "reset": true, 00:24:37.562 "nvme_admin": false, 00:24:37.562 "nvme_io": false, 00:24:37.562 "nvme_io_md": false, 00:24:37.562 "write_zeroes": true, 00:24:37.562 "zcopy": true, 00:24:37.562 "get_zone_info": false, 00:24:37.562 "zone_management": false, 00:24:37.562 "zone_append": false, 00:24:37.562 "compare": false, 00:24:37.562 "compare_and_write": false, 00:24:37.562 "abort": true, 00:24:37.562 "seek_hole": false, 00:24:37.562 "seek_data": false, 00:24:37.562 "copy": true, 00:24:37.562 "nvme_iov_md": false 00:24:37.562 }, 00:24:37.562 "memory_domains": [ 00:24:37.562 { 00:24:37.562 "dma_device_id": "system", 00:24:37.562 "dma_device_type": 1 00:24:37.562 }, 00:24:37.562 { 00:24:37.562 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:37.562 "dma_device_type": 2 00:24:37.562 } 00:24:37.562 ], 00:24:37.562 "driver_specific": {} 00:24:37.562 } 00:24:37.562 ]' 00:24:37.562 13:44:44 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # jq length 00:24:37.562 13:44:44 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:24:37.562 13:44:44 rpc.rpc_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc0 -p Passthru0 00:24:37.562 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:37.562 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:24:37.562 [2024-11-20 13:44:44.516696] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc0 00:24:37.562 [2024-11-20 13:44:44.516763] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:37.562 [2024-11-20 13:44:44.516784] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:24:37.562 [2024-11-20 13:44:44.516799] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:37.562 [2024-11-20 13:44:44.519014] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:37.562 [2024-11-20 13:44:44.519066] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:24:37.562 Passthru0 00:24:37.562 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:37.562 13:44:44 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:24:37.562 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:37.562 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:24:37.562 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:37.562 13:44:44 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:24:37.562 { 00:24:37.562 "name": "Malloc0", 00:24:37.562 "aliases": [ 00:24:37.562 "97a267fa-718c-4c9a-a372-614170c633f1" 00:24:37.562 ], 00:24:37.562 "product_name": "Malloc disk", 00:24:37.562 "block_size": 512, 00:24:37.562 "num_blocks": 16384, 00:24:37.562 "uuid": "97a267fa-718c-4c9a-a372-614170c633f1", 00:24:37.562 "assigned_rate_limits": { 00:24:37.562 "rw_ios_per_sec": 0, 00:24:37.562 "rw_mbytes_per_sec": 0, 00:24:37.562 "r_mbytes_per_sec": 0, 00:24:37.562 "w_mbytes_per_sec": 0 00:24:37.562 }, 00:24:37.562 "claimed": true, 00:24:37.562 "claim_type": "exclusive_write", 00:24:37.562 "zoned": false, 00:24:37.562 "supported_io_types": { 00:24:37.562 "read": true, 00:24:37.562 "write": true, 00:24:37.562 "unmap": true, 00:24:37.562 "flush": true, 00:24:37.562 "reset": true, 00:24:37.562 "nvme_admin": false, 00:24:37.562 "nvme_io": false, 00:24:37.562 "nvme_io_md": false, 00:24:37.562 "write_zeroes": true, 00:24:37.562 "zcopy": true, 00:24:37.562 "get_zone_info": false, 00:24:37.563 "zone_management": false, 00:24:37.563 "zone_append": false, 00:24:37.563 "compare": false, 00:24:37.563 "compare_and_write": false, 00:24:37.563 "abort": true, 00:24:37.563 "seek_hole": false, 00:24:37.563 "seek_data": false, 00:24:37.563 "copy": true, 00:24:37.563 "nvme_iov_md": false 00:24:37.563 }, 00:24:37.563 "memory_domains": [ 00:24:37.563 { 00:24:37.563 "dma_device_id": "system", 00:24:37.563 "dma_device_type": 1 00:24:37.563 }, 00:24:37.563 { 00:24:37.563 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:37.563 "dma_device_type": 2 00:24:37.563 } 00:24:37.563 ], 00:24:37.563 "driver_specific": {} 00:24:37.563 }, 00:24:37.563 { 00:24:37.563 "name": "Passthru0", 00:24:37.563 "aliases": [ 00:24:37.563 "20a40e4c-4a2c-5909-a0c7-41572e7d14e9" 00:24:37.563 ], 00:24:37.563 "product_name": "passthru", 00:24:37.563 "block_size": 512, 00:24:37.563 "num_blocks": 16384, 00:24:37.563 "uuid": "20a40e4c-4a2c-5909-a0c7-41572e7d14e9", 00:24:37.563 "assigned_rate_limits": { 00:24:37.563 "rw_ios_per_sec": 0, 00:24:37.563 "rw_mbytes_per_sec": 0, 00:24:37.563 "r_mbytes_per_sec": 0, 00:24:37.563 "w_mbytes_per_sec": 0 00:24:37.563 }, 00:24:37.563 "claimed": false, 00:24:37.563 "zoned": false, 00:24:37.563 "supported_io_types": { 00:24:37.563 "read": true, 00:24:37.563 "write": true, 00:24:37.563 "unmap": true, 00:24:37.563 "flush": true, 00:24:37.563 "reset": true, 00:24:37.563 "nvme_admin": false, 00:24:37.563 "nvme_io": false, 00:24:37.563 "nvme_io_md": false, 00:24:37.563 "write_zeroes": true, 00:24:37.563 "zcopy": true, 00:24:37.563 "get_zone_info": false, 00:24:37.563 "zone_management": false, 00:24:37.563 "zone_append": false, 00:24:37.563 "compare": false, 00:24:37.563 "compare_and_write": false, 00:24:37.563 "abort": true, 00:24:37.563 "seek_hole": false, 00:24:37.563 "seek_data": false, 00:24:37.563 "copy": true, 00:24:37.563 "nvme_iov_md": false 00:24:37.563 }, 00:24:37.563 "memory_domains": [ 00:24:37.563 { 00:24:37.563 "dma_device_id": "system", 00:24:37.563 "dma_device_type": 1 00:24:37.563 }, 00:24:37.563 { 00:24:37.563 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:37.563 "dma_device_type": 2 00:24:37.563 } 00:24:37.563 ], 00:24:37.563 "driver_specific": { 00:24:37.563 "passthru": { 00:24:37.563 "name": "Passthru0", 00:24:37.563 "base_bdev_name": "Malloc0" 00:24:37.563 } 00:24:37.563 } 00:24:37.563 } 00:24:37.563 ]' 00:24:37.563 13:44:44 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # jq length 00:24:37.563 13:44:44 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:24:37.563 13:44:44 rpc.rpc_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:24:37.563 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:37.563 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:24:37.563 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:37.563 13:44:44 rpc.rpc_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc0 00:24:37.563 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:37.563 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:24:37.563 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:37.563 13:44:44 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:24:37.563 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:37.563 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:24:37.563 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:37.563 13:44:44 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:24:37.563 13:44:44 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # jq length 00:24:37.821 13:44:44 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:24:37.821 00:24:37.821 real 0m0.244s 00:24:37.821 user 0m0.129s 00:24:37.821 sys 0m0.033s 00:24:37.821 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@1130 -- # xtrace_disable 00:24:37.821 13:44:44 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:24:37.821 ************************************ 00:24:37.821 END TEST rpc_integrity 00:24:37.821 ************************************ 00:24:37.821 13:44:44 rpc -- rpc/rpc.sh@74 -- # run_test rpc_plugins rpc_plugins 00:24:37.821 13:44:44 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:24:37.821 13:44:44 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:24:37.821 13:44:44 rpc -- common/autotest_common.sh@10 -- # set +x 00:24:37.821 ************************************ 00:24:37.821 START TEST rpc_plugins 00:24:37.821 ************************************ 00:24:37.821 13:44:44 rpc.rpc_plugins -- common/autotest_common.sh@1129 -- # rpc_plugins 00:24:37.821 13:44:44 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # rpc_cmd --plugin rpc_plugin create_malloc 00:24:37.821 13:44:44 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:37.821 13:44:44 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:24:37.821 13:44:44 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:37.821 13:44:44 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # malloc=Malloc1 00:24:37.821 13:44:44 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # rpc_cmd bdev_get_bdevs 00:24:37.821 13:44:44 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:37.821 13:44:44 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:24:37.821 13:44:44 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:37.821 13:44:44 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # bdevs='[ 00:24:37.821 { 00:24:37.821 "name": "Malloc1", 00:24:37.821 "aliases": [ 00:24:37.821 "578ccf68-7dc9-4707-959c-3bbb72cc028d" 00:24:37.821 ], 00:24:37.821 "product_name": "Malloc disk", 00:24:37.821 "block_size": 4096, 00:24:37.821 "num_blocks": 256, 00:24:37.821 "uuid": "578ccf68-7dc9-4707-959c-3bbb72cc028d", 00:24:37.821 "assigned_rate_limits": { 00:24:37.821 "rw_ios_per_sec": 0, 00:24:37.821 "rw_mbytes_per_sec": 0, 00:24:37.821 "r_mbytes_per_sec": 0, 00:24:37.821 "w_mbytes_per_sec": 0 00:24:37.821 }, 00:24:37.821 "claimed": false, 00:24:37.821 "zoned": false, 00:24:37.821 "supported_io_types": { 00:24:37.821 "read": true, 00:24:37.821 "write": true, 00:24:37.821 "unmap": true, 00:24:37.821 "flush": true, 00:24:37.821 "reset": true, 00:24:37.821 "nvme_admin": false, 00:24:37.821 "nvme_io": false, 00:24:37.821 "nvme_io_md": false, 00:24:37.821 "write_zeroes": true, 00:24:37.821 "zcopy": true, 00:24:37.821 "get_zone_info": false, 00:24:37.821 "zone_management": false, 00:24:37.821 "zone_append": false, 00:24:37.821 "compare": false, 00:24:37.821 "compare_and_write": false, 00:24:37.821 "abort": true, 00:24:37.821 "seek_hole": false, 00:24:37.821 "seek_data": false, 00:24:37.821 "copy": true, 00:24:37.821 "nvme_iov_md": false 00:24:37.821 }, 00:24:37.821 "memory_domains": [ 00:24:37.821 { 00:24:37.821 "dma_device_id": "system", 00:24:37.821 "dma_device_type": 1 00:24:37.821 }, 00:24:37.821 { 00:24:37.821 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:37.821 "dma_device_type": 2 00:24:37.821 } 00:24:37.821 ], 00:24:37.821 "driver_specific": {} 00:24:37.821 } 00:24:37.821 ]' 00:24:37.821 13:44:44 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # jq length 00:24:37.821 13:44:44 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # '[' 1 == 1 ']' 00:24:37.821 13:44:44 rpc.rpc_plugins -- rpc/rpc.sh@34 -- # rpc_cmd --plugin rpc_plugin delete_malloc Malloc1 00:24:37.821 13:44:44 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:37.821 13:44:44 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:24:37.822 13:44:44 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:37.822 13:44:44 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # rpc_cmd bdev_get_bdevs 00:24:37.822 13:44:44 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:37.822 13:44:44 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:24:37.822 13:44:44 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:37.822 13:44:44 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # bdevs='[]' 00:24:37.822 13:44:44 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # jq length 00:24:37.822 13:44:44 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # '[' 0 == 0 ']' 00:24:37.822 00:24:37.822 real 0m0.113s 00:24:37.822 user 0m0.060s 00:24:37.822 sys 0m0.018s 00:24:37.822 13:44:44 rpc.rpc_plugins -- common/autotest_common.sh@1130 -- # xtrace_disable 00:24:37.822 13:44:44 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:24:37.822 ************************************ 00:24:37.822 END TEST rpc_plugins 00:24:37.822 ************************************ 00:24:37.822 13:44:44 rpc -- rpc/rpc.sh@75 -- # run_test rpc_trace_cmd_test rpc_trace_cmd_test 00:24:37.822 13:44:44 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:24:37.822 13:44:44 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:24:37.822 13:44:44 rpc -- common/autotest_common.sh@10 -- # set +x 00:24:37.822 ************************************ 00:24:37.822 START TEST rpc_trace_cmd_test 00:24:37.822 ************************************ 00:24:37.822 13:44:44 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1129 -- # rpc_trace_cmd_test 00:24:37.822 13:44:44 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@40 -- # local info 00:24:37.822 13:44:44 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # rpc_cmd trace_get_info 00:24:37.822 13:44:44 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:37.822 13:44:44 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:24:37.822 13:44:44 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:37.822 13:44:44 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # info='{ 00:24:37.822 "tpoint_shm_path": "/dev/shm/spdk_tgt_trace.pid56103", 00:24:37.822 "tpoint_group_mask": "0x8", 00:24:37.822 "iscsi_conn": { 00:24:37.822 "mask": "0x2", 00:24:37.822 "tpoint_mask": "0x0" 00:24:37.822 }, 00:24:37.822 "scsi": { 00:24:37.822 "mask": "0x4", 00:24:37.822 "tpoint_mask": "0x0" 00:24:37.822 }, 00:24:37.822 "bdev": { 00:24:37.822 "mask": "0x8", 00:24:37.822 "tpoint_mask": "0xffffffffffffffff" 00:24:37.822 }, 00:24:37.822 "nvmf_rdma": { 00:24:37.822 "mask": "0x10", 00:24:37.822 "tpoint_mask": "0x0" 00:24:37.822 }, 00:24:37.822 "nvmf_tcp": { 00:24:37.822 "mask": "0x20", 00:24:37.822 "tpoint_mask": "0x0" 00:24:37.822 }, 00:24:37.822 "ftl": { 00:24:37.822 "mask": "0x40", 00:24:37.822 "tpoint_mask": "0x0" 00:24:37.822 }, 00:24:37.822 "blobfs": { 00:24:37.822 "mask": "0x80", 00:24:37.822 "tpoint_mask": "0x0" 00:24:37.822 }, 00:24:37.822 "dsa": { 00:24:37.822 "mask": "0x200", 00:24:37.822 "tpoint_mask": "0x0" 00:24:37.822 }, 00:24:37.822 "thread": { 00:24:37.822 "mask": "0x400", 00:24:37.822 "tpoint_mask": "0x0" 00:24:37.822 }, 00:24:37.822 "nvme_pcie": { 00:24:37.822 "mask": "0x800", 00:24:37.822 "tpoint_mask": "0x0" 00:24:37.822 }, 00:24:37.822 "iaa": { 00:24:37.822 "mask": "0x1000", 00:24:37.822 "tpoint_mask": "0x0" 00:24:37.822 }, 00:24:37.822 "nvme_tcp": { 00:24:37.822 "mask": "0x2000", 00:24:37.822 "tpoint_mask": "0x0" 00:24:37.822 }, 00:24:37.822 "bdev_nvme": { 00:24:37.822 "mask": "0x4000", 00:24:37.822 "tpoint_mask": "0x0" 00:24:37.822 }, 00:24:37.822 "sock": { 00:24:37.822 "mask": "0x8000", 00:24:37.822 "tpoint_mask": "0x0" 00:24:37.822 }, 00:24:37.822 "blob": { 00:24:37.822 "mask": "0x10000", 00:24:37.822 "tpoint_mask": "0x0" 00:24:37.822 }, 00:24:37.822 "bdev_raid": { 00:24:37.822 "mask": "0x20000", 00:24:37.822 "tpoint_mask": "0x0" 00:24:37.822 }, 00:24:37.822 "scheduler": { 00:24:37.822 "mask": "0x40000", 00:24:37.822 "tpoint_mask": "0x0" 00:24:37.822 } 00:24:37.822 }' 00:24:37.822 13:44:44 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # jq length 00:24:38.080 13:44:44 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # '[' 19 -gt 2 ']' 00:24:38.080 13:44:44 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # jq 'has("tpoint_group_mask")' 00:24:38.080 13:44:44 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # '[' true = true ']' 00:24:38.080 13:44:44 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # jq 'has("tpoint_shm_path")' 00:24:38.080 13:44:44 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # '[' true = true ']' 00:24:38.080 13:44:44 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # jq 'has("bdev")' 00:24:38.080 13:44:44 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # '[' true = true ']' 00:24:38.080 13:44:44 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # jq -r .bdev.tpoint_mask 00:24:38.080 13:44:45 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # '[' 0xffffffffffffffff '!=' 0x0 ']' 00:24:38.080 00:24:38.080 real 0m0.180s 00:24:38.080 user 0m0.144s 00:24:38.080 sys 0m0.027s 00:24:38.080 ************************************ 00:24:38.080 END TEST rpc_trace_cmd_test 00:24:38.080 ************************************ 00:24:38.080 13:44:45 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:24:38.080 13:44:45 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:24:38.080 13:44:45 rpc -- rpc/rpc.sh@76 -- # [[ 0 -eq 1 ]] 00:24:38.080 13:44:45 rpc -- rpc/rpc.sh@80 -- # rpc=rpc_cmd 00:24:38.080 13:44:45 rpc -- rpc/rpc.sh@81 -- # run_test rpc_daemon_integrity rpc_integrity 00:24:38.080 13:44:45 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:24:38.080 13:44:45 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:24:38.080 13:44:45 rpc -- common/autotest_common.sh@10 -- # set +x 00:24:38.080 ************************************ 00:24:38.080 START TEST rpc_daemon_integrity 00:24:38.080 ************************************ 00:24:38.080 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1129 -- # rpc_integrity 00:24:38.080 13:44:45 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:24:38.080 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:38.080 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:24:38.080 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:38.080 13:44:45 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:24:38.080 13:44:45 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # jq length 00:24:38.080 13:44:45 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:24:38.081 13:44:45 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:24:38.081 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:38.081 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:24:38.081 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:38.081 13:44:45 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc2 00:24:38.081 13:44:45 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:24:38.081 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:38.081 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:24:38.339 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:38.339 13:44:45 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:24:38.339 { 00:24:38.339 "name": "Malloc2", 00:24:38.339 "aliases": [ 00:24:38.339 "a49ddac1-4d7c-4cac-bddd-4bc12e60fdc2" 00:24:38.339 ], 00:24:38.339 "product_name": "Malloc disk", 00:24:38.339 "block_size": 512, 00:24:38.339 "num_blocks": 16384, 00:24:38.339 "uuid": "a49ddac1-4d7c-4cac-bddd-4bc12e60fdc2", 00:24:38.339 "assigned_rate_limits": { 00:24:38.339 "rw_ios_per_sec": 0, 00:24:38.339 "rw_mbytes_per_sec": 0, 00:24:38.339 "r_mbytes_per_sec": 0, 00:24:38.339 "w_mbytes_per_sec": 0 00:24:38.340 }, 00:24:38.340 "claimed": false, 00:24:38.340 "zoned": false, 00:24:38.340 "supported_io_types": { 00:24:38.340 "read": true, 00:24:38.340 "write": true, 00:24:38.340 "unmap": true, 00:24:38.340 "flush": true, 00:24:38.340 "reset": true, 00:24:38.340 "nvme_admin": false, 00:24:38.340 "nvme_io": false, 00:24:38.340 "nvme_io_md": false, 00:24:38.340 "write_zeroes": true, 00:24:38.340 "zcopy": true, 00:24:38.340 "get_zone_info": false, 00:24:38.340 "zone_management": false, 00:24:38.340 "zone_append": false, 00:24:38.340 "compare": false, 00:24:38.340 "compare_and_write": false, 00:24:38.340 "abort": true, 00:24:38.340 "seek_hole": false, 00:24:38.340 "seek_data": false, 00:24:38.340 "copy": true, 00:24:38.340 "nvme_iov_md": false 00:24:38.340 }, 00:24:38.340 "memory_domains": [ 00:24:38.340 { 00:24:38.340 "dma_device_id": "system", 00:24:38.340 "dma_device_type": 1 00:24:38.340 }, 00:24:38.340 { 00:24:38.340 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:38.340 "dma_device_type": 2 00:24:38.340 } 00:24:38.340 ], 00:24:38.340 "driver_specific": {} 00:24:38.340 } 00:24:38.340 ]' 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # jq length 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc2 -p Passthru0 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:24:38.340 [2024-11-20 13:44:45.172392] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc2 00:24:38.340 [2024-11-20 13:44:45.172461] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:38.340 [2024-11-20 13:44:45.172482] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:24:38.340 [2024-11-20 13:44:45.172493] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:38.340 [2024-11-20 13:44:45.174671] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:38.340 [2024-11-20 13:44:45.174713] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:24:38.340 Passthru0 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:24:38.340 { 00:24:38.340 "name": "Malloc2", 00:24:38.340 "aliases": [ 00:24:38.340 "a49ddac1-4d7c-4cac-bddd-4bc12e60fdc2" 00:24:38.340 ], 00:24:38.340 "product_name": "Malloc disk", 00:24:38.340 "block_size": 512, 00:24:38.340 "num_blocks": 16384, 00:24:38.340 "uuid": "a49ddac1-4d7c-4cac-bddd-4bc12e60fdc2", 00:24:38.340 "assigned_rate_limits": { 00:24:38.340 "rw_ios_per_sec": 0, 00:24:38.340 "rw_mbytes_per_sec": 0, 00:24:38.340 "r_mbytes_per_sec": 0, 00:24:38.340 "w_mbytes_per_sec": 0 00:24:38.340 }, 00:24:38.340 "claimed": true, 00:24:38.340 "claim_type": "exclusive_write", 00:24:38.340 "zoned": false, 00:24:38.340 "supported_io_types": { 00:24:38.340 "read": true, 00:24:38.340 "write": true, 00:24:38.340 "unmap": true, 00:24:38.340 "flush": true, 00:24:38.340 "reset": true, 00:24:38.340 "nvme_admin": false, 00:24:38.340 "nvme_io": false, 00:24:38.340 "nvme_io_md": false, 00:24:38.340 "write_zeroes": true, 00:24:38.340 "zcopy": true, 00:24:38.340 "get_zone_info": false, 00:24:38.340 "zone_management": false, 00:24:38.340 "zone_append": false, 00:24:38.340 "compare": false, 00:24:38.340 "compare_and_write": false, 00:24:38.340 "abort": true, 00:24:38.340 "seek_hole": false, 00:24:38.340 "seek_data": false, 00:24:38.340 "copy": true, 00:24:38.340 "nvme_iov_md": false 00:24:38.340 }, 00:24:38.340 "memory_domains": [ 00:24:38.340 { 00:24:38.340 "dma_device_id": "system", 00:24:38.340 "dma_device_type": 1 00:24:38.340 }, 00:24:38.340 { 00:24:38.340 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:38.340 "dma_device_type": 2 00:24:38.340 } 00:24:38.340 ], 00:24:38.340 "driver_specific": {} 00:24:38.340 }, 00:24:38.340 { 00:24:38.340 "name": "Passthru0", 00:24:38.340 "aliases": [ 00:24:38.340 "d3e329d1-fc95-51b4-a7b5-6e528061cbc9" 00:24:38.340 ], 00:24:38.340 "product_name": "passthru", 00:24:38.340 "block_size": 512, 00:24:38.340 "num_blocks": 16384, 00:24:38.340 "uuid": "d3e329d1-fc95-51b4-a7b5-6e528061cbc9", 00:24:38.340 "assigned_rate_limits": { 00:24:38.340 "rw_ios_per_sec": 0, 00:24:38.340 "rw_mbytes_per_sec": 0, 00:24:38.340 "r_mbytes_per_sec": 0, 00:24:38.340 "w_mbytes_per_sec": 0 00:24:38.340 }, 00:24:38.340 "claimed": false, 00:24:38.340 "zoned": false, 00:24:38.340 "supported_io_types": { 00:24:38.340 "read": true, 00:24:38.340 "write": true, 00:24:38.340 "unmap": true, 00:24:38.340 "flush": true, 00:24:38.340 "reset": true, 00:24:38.340 "nvme_admin": false, 00:24:38.340 "nvme_io": false, 00:24:38.340 "nvme_io_md": false, 00:24:38.340 "write_zeroes": true, 00:24:38.340 "zcopy": true, 00:24:38.340 "get_zone_info": false, 00:24:38.340 "zone_management": false, 00:24:38.340 "zone_append": false, 00:24:38.340 "compare": false, 00:24:38.340 "compare_and_write": false, 00:24:38.340 "abort": true, 00:24:38.340 "seek_hole": false, 00:24:38.340 "seek_data": false, 00:24:38.340 "copy": true, 00:24:38.340 "nvme_iov_md": false 00:24:38.340 }, 00:24:38.340 "memory_domains": [ 00:24:38.340 { 00:24:38.340 "dma_device_id": "system", 00:24:38.340 "dma_device_type": 1 00:24:38.340 }, 00:24:38.340 { 00:24:38.340 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:38.340 "dma_device_type": 2 00:24:38.340 } 00:24:38.340 ], 00:24:38.340 "driver_specific": { 00:24:38.340 "passthru": { 00:24:38.340 "name": "Passthru0", 00:24:38.340 "base_bdev_name": "Malloc2" 00:24:38.340 } 00:24:38.340 } 00:24:38.340 } 00:24:38.340 ]' 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # jq length 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc2 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # jq length 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:24:38.340 00:24:38.340 real 0m0.233s 00:24:38.340 user 0m0.122s 00:24:38.340 sys 0m0.037s 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1130 -- # xtrace_disable 00:24:38.340 13:44:45 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:24:38.340 ************************************ 00:24:38.340 END TEST rpc_daemon_integrity 00:24:38.340 ************************************ 00:24:38.340 13:44:45 rpc -- rpc/rpc.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:24:38.340 13:44:45 rpc -- rpc/rpc.sh@84 -- # killprocess 56103 00:24:38.341 13:44:45 rpc -- common/autotest_common.sh@954 -- # '[' -z 56103 ']' 00:24:38.341 13:44:45 rpc -- common/autotest_common.sh@958 -- # kill -0 56103 00:24:38.341 13:44:45 rpc -- common/autotest_common.sh@959 -- # uname 00:24:38.341 13:44:45 rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:24:38.341 13:44:45 rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 56103 00:24:38.341 13:44:45 rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:24:38.341 13:44:45 rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:24:38.341 killing process with pid 56103 00:24:38.341 13:44:45 rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 56103' 00:24:38.341 13:44:45 rpc -- common/autotest_common.sh@973 -- # kill 56103 00:24:38.341 13:44:45 rpc -- common/autotest_common.sh@978 -- # wait 56103 00:24:40.243 00:24:40.243 real 0m3.700s 00:24:40.243 user 0m4.090s 00:24:40.243 sys 0m0.641s 00:24:40.243 13:44:46 rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:24:40.243 ************************************ 00:24:40.243 END TEST rpc 00:24:40.243 ************************************ 00:24:40.243 13:44:46 rpc -- common/autotest_common.sh@10 -- # set +x 00:24:40.243 13:44:47 -- spdk/autotest.sh@157 -- # run_test skip_rpc /home/vagrant/spdk_repo/spdk/test/rpc/skip_rpc.sh 00:24:40.243 13:44:47 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:24:40.243 13:44:47 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:24:40.243 13:44:47 -- common/autotest_common.sh@10 -- # set +x 00:24:40.243 ************************************ 00:24:40.243 START TEST skip_rpc 00:24:40.243 ************************************ 00:24:40.243 13:44:47 skip_rpc -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/rpc/skip_rpc.sh 00:24:40.243 * Looking for test storage... 00:24:40.243 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc 00:24:40.243 13:44:47 skip_rpc -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:24:40.243 13:44:47 skip_rpc -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:24:40.243 13:44:47 skip_rpc -- common/autotest_common.sh@1693 -- # lcov --version 00:24:40.243 13:44:47 skip_rpc -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@344 -- # case "$op" in 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@345 -- # : 1 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@365 -- # decimal 1 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@353 -- # local d=1 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@355 -- # echo 1 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@366 -- # decimal 2 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@353 -- # local d=2 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@355 -- # echo 2 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:24:40.243 13:44:47 skip_rpc -- scripts/common.sh@368 -- # return 0 00:24:40.243 13:44:47 skip_rpc -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:24:40.243 13:44:47 skip_rpc -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:24:40.243 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:40.243 --rc genhtml_branch_coverage=1 00:24:40.243 --rc genhtml_function_coverage=1 00:24:40.243 --rc genhtml_legend=1 00:24:40.243 --rc geninfo_all_blocks=1 00:24:40.243 --rc geninfo_unexecuted_blocks=1 00:24:40.243 00:24:40.243 ' 00:24:40.243 13:44:47 skip_rpc -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:24:40.243 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:40.243 --rc genhtml_branch_coverage=1 00:24:40.243 --rc genhtml_function_coverage=1 00:24:40.243 --rc genhtml_legend=1 00:24:40.243 --rc geninfo_all_blocks=1 00:24:40.243 --rc geninfo_unexecuted_blocks=1 00:24:40.243 00:24:40.243 ' 00:24:40.243 13:44:47 skip_rpc -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:24:40.243 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:40.243 --rc genhtml_branch_coverage=1 00:24:40.243 --rc genhtml_function_coverage=1 00:24:40.243 --rc genhtml_legend=1 00:24:40.243 --rc geninfo_all_blocks=1 00:24:40.243 --rc geninfo_unexecuted_blocks=1 00:24:40.243 00:24:40.243 ' 00:24:40.243 13:44:47 skip_rpc -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:24:40.243 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:40.243 --rc genhtml_branch_coverage=1 00:24:40.243 --rc genhtml_function_coverage=1 00:24:40.243 --rc genhtml_legend=1 00:24:40.243 --rc geninfo_all_blocks=1 00:24:40.243 --rc geninfo_unexecuted_blocks=1 00:24:40.243 00:24:40.243 ' 00:24:40.243 13:44:47 skip_rpc -- rpc/skip_rpc.sh@11 -- # CONFIG_PATH=/home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:24:40.243 13:44:47 skip_rpc -- rpc/skip_rpc.sh@12 -- # LOG_PATH=/home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:24:40.243 13:44:47 skip_rpc -- rpc/skip_rpc.sh@73 -- # run_test skip_rpc test_skip_rpc 00:24:40.243 13:44:47 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:24:40.243 13:44:47 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:24:40.243 13:44:47 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:24:40.243 ************************************ 00:24:40.243 START TEST skip_rpc 00:24:40.243 ************************************ 00:24:40.243 13:44:47 skip_rpc.skip_rpc -- common/autotest_common.sh@1129 -- # test_skip_rpc 00:24:40.243 13:44:47 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@16 -- # local spdk_pid=56321 00:24:40.243 13:44:47 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@18 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:24:40.243 13:44:47 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@19 -- # sleep 5 00:24:40.243 13:44:47 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@15 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 00:24:40.243 [2024-11-20 13:44:47.240936] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:24:40.243 [2024-11-20 13:44:47.241107] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid56321 ] 00:24:40.501 [2024-11-20 13:44:47.408915] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:40.501 [2024-11-20 13:44:47.524557] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@21 -- # NOT rpc_cmd spdk_get_version 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- common/autotest_common.sh@652 -- # local es=0 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd spdk_get_version 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- common/autotest_common.sh@655 -- # rpc_cmd spdk_get_version 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- common/autotest_common.sh@655 -- # es=1 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@22 -- # trap - SIGINT SIGTERM EXIT 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@23 -- # killprocess 56321 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- common/autotest_common.sh@954 -- # '[' -z 56321 ']' 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- common/autotest_common.sh@958 -- # kill -0 56321 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- common/autotest_common.sh@959 -- # uname 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 56321 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:24:45.761 killing process with pid 56321 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 56321' 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- common/autotest_common.sh@973 -- # kill 56321 00:24:45.761 13:44:52 skip_rpc.skip_rpc -- common/autotest_common.sh@978 -- # wait 56321 00:24:46.696 00:24:46.696 real 0m6.328s 00:24:46.696 user 0m5.885s 00:24:46.696 sys 0m0.337s 00:24:46.696 13:44:53 skip_rpc.skip_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:24:46.696 13:44:53 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:24:46.696 ************************************ 00:24:46.696 END TEST skip_rpc 00:24:46.696 ************************************ 00:24:46.696 13:44:53 skip_rpc -- rpc/skip_rpc.sh@74 -- # run_test skip_rpc_with_json test_skip_rpc_with_json 00:24:46.696 13:44:53 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:24:46.696 13:44:53 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:24:46.696 13:44:53 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:24:46.696 ************************************ 00:24:46.696 START TEST skip_rpc_with_json 00:24:46.696 ************************************ 00:24:46.696 13:44:53 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1129 -- # test_skip_rpc_with_json 00:24:46.696 13:44:53 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@44 -- # gen_json_config 00:24:46.696 13:44:53 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@28 -- # local spdk_pid=56414 00:24:46.696 13:44:53 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@30 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:24:46.696 13:44:53 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@31 -- # waitforlisten 56414 00:24:46.696 13:44:53 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@835 -- # '[' -z 56414 ']' 00:24:46.696 13:44:53 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:46.696 13:44:53 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@840 -- # local max_retries=100 00:24:46.696 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:46.696 13:44:53 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:46.696 13:44:53 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@844 -- # xtrace_disable 00:24:46.696 13:44:53 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:24:46.696 13:44:53 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@27 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:24:46.696 [2024-11-20 13:44:53.609963] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:24:46.696 [2024-11-20 13:44:53.610110] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid56414 ] 00:24:46.954 [2024-11-20 13:44:53.766500] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:46.954 [2024-11-20 13:44:53.867997] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:24:47.521 13:44:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:24:47.521 13:44:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@868 -- # return 0 00:24:47.521 13:44:54 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_get_transports --trtype tcp 00:24:47.521 13:44:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:47.521 13:44:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:24:47.521 [2024-11-20 13:44:54.415345] nvmf_rpc.c:2706:rpc_nvmf_get_transports: *ERROR*: transport 'tcp' does not exist 00:24:47.521 request: 00:24:47.521 { 00:24:47.521 "trtype": "tcp", 00:24:47.521 "method": "nvmf_get_transports", 00:24:47.521 "req_id": 1 00:24:47.521 } 00:24:47.521 Got JSON-RPC error response 00:24:47.521 response: 00:24:47.521 { 00:24:47.521 "code": -19, 00:24:47.521 "message": "No such device" 00:24:47.521 } 00:24:47.521 13:44:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:24:47.521 13:44:54 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_create_transport -t tcp 00:24:47.521 13:44:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:47.521 13:44:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:24:47.521 [2024-11-20 13:44:54.423434] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:24:47.521 13:44:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:47.521 13:44:54 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@36 -- # rpc_cmd save_config 00:24:47.521 13:44:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:47.521 13:44:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:24:47.778 13:44:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:47.778 13:44:54 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@37 -- # cat /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:24:47.778 { 00:24:47.778 "subsystems": [ 00:24:47.778 { 00:24:47.778 "subsystem": "fsdev", 00:24:47.778 "config": [ 00:24:47.778 { 00:24:47.778 "method": "fsdev_set_opts", 00:24:47.778 "params": { 00:24:47.778 "fsdev_io_pool_size": 65535, 00:24:47.778 "fsdev_io_cache_size": 256 00:24:47.778 } 00:24:47.778 } 00:24:47.778 ] 00:24:47.778 }, 00:24:47.778 { 00:24:47.778 "subsystem": "keyring", 00:24:47.778 "config": [] 00:24:47.778 }, 00:24:47.778 { 00:24:47.778 "subsystem": "iobuf", 00:24:47.778 "config": [ 00:24:47.778 { 00:24:47.778 "method": "iobuf_set_options", 00:24:47.778 "params": { 00:24:47.778 "small_pool_count": 8192, 00:24:47.778 "large_pool_count": 1024, 00:24:47.778 "small_bufsize": 8192, 00:24:47.778 "large_bufsize": 135168, 00:24:47.778 "enable_numa": false 00:24:47.778 } 00:24:47.778 } 00:24:47.778 ] 00:24:47.778 }, 00:24:47.778 { 00:24:47.778 "subsystem": "sock", 00:24:47.778 "config": [ 00:24:47.778 { 00:24:47.778 "method": "sock_set_default_impl", 00:24:47.778 "params": { 00:24:47.778 "impl_name": "posix" 00:24:47.778 } 00:24:47.778 }, 00:24:47.778 { 00:24:47.778 "method": "sock_impl_set_options", 00:24:47.778 "params": { 00:24:47.778 "impl_name": "ssl", 00:24:47.778 "recv_buf_size": 4096, 00:24:47.778 "send_buf_size": 4096, 00:24:47.778 "enable_recv_pipe": true, 00:24:47.778 "enable_quickack": false, 00:24:47.778 "enable_placement_id": 0, 00:24:47.778 "enable_zerocopy_send_server": true, 00:24:47.778 "enable_zerocopy_send_client": false, 00:24:47.778 "zerocopy_threshold": 0, 00:24:47.778 "tls_version": 0, 00:24:47.778 "enable_ktls": false 00:24:47.778 } 00:24:47.778 }, 00:24:47.778 { 00:24:47.778 "method": "sock_impl_set_options", 00:24:47.778 "params": { 00:24:47.778 "impl_name": "posix", 00:24:47.778 "recv_buf_size": 2097152, 00:24:47.778 "send_buf_size": 2097152, 00:24:47.778 "enable_recv_pipe": true, 00:24:47.778 "enable_quickack": false, 00:24:47.778 "enable_placement_id": 0, 00:24:47.778 "enable_zerocopy_send_server": true, 00:24:47.778 "enable_zerocopy_send_client": false, 00:24:47.778 "zerocopy_threshold": 0, 00:24:47.778 "tls_version": 0, 00:24:47.778 "enable_ktls": false 00:24:47.778 } 00:24:47.778 } 00:24:47.778 ] 00:24:47.778 }, 00:24:47.778 { 00:24:47.778 "subsystem": "vmd", 00:24:47.778 "config": [] 00:24:47.778 }, 00:24:47.778 { 00:24:47.778 "subsystem": "accel", 00:24:47.778 "config": [ 00:24:47.778 { 00:24:47.778 "method": "accel_set_options", 00:24:47.778 "params": { 00:24:47.779 "small_cache_size": 128, 00:24:47.779 "large_cache_size": 16, 00:24:47.779 "task_count": 2048, 00:24:47.779 "sequence_count": 2048, 00:24:47.779 "buf_count": 2048 00:24:47.779 } 00:24:47.779 } 00:24:47.779 ] 00:24:47.779 }, 00:24:47.779 { 00:24:47.779 "subsystem": "bdev", 00:24:47.779 "config": [ 00:24:47.779 { 00:24:47.779 "method": "bdev_set_options", 00:24:47.779 "params": { 00:24:47.779 "bdev_io_pool_size": 65535, 00:24:47.779 "bdev_io_cache_size": 256, 00:24:47.779 "bdev_auto_examine": true, 00:24:47.779 "iobuf_small_cache_size": 128, 00:24:47.779 "iobuf_large_cache_size": 16 00:24:47.779 } 00:24:47.779 }, 00:24:47.779 { 00:24:47.779 "method": "bdev_raid_set_options", 00:24:47.779 "params": { 00:24:47.779 "process_window_size_kb": 1024, 00:24:47.779 "process_max_bandwidth_mb_sec": 0 00:24:47.779 } 00:24:47.779 }, 00:24:47.779 { 00:24:47.779 "method": "bdev_iscsi_set_options", 00:24:47.779 "params": { 00:24:47.779 "timeout_sec": 30 00:24:47.779 } 00:24:47.779 }, 00:24:47.779 { 00:24:47.779 "method": "bdev_nvme_set_options", 00:24:47.779 "params": { 00:24:47.779 "action_on_timeout": "none", 00:24:47.779 "timeout_us": 0, 00:24:47.779 "timeout_admin_us": 0, 00:24:47.779 "keep_alive_timeout_ms": 10000, 00:24:47.779 "arbitration_burst": 0, 00:24:47.779 "low_priority_weight": 0, 00:24:47.779 "medium_priority_weight": 0, 00:24:47.779 "high_priority_weight": 0, 00:24:47.779 "nvme_adminq_poll_period_us": 10000, 00:24:47.779 "nvme_ioq_poll_period_us": 0, 00:24:47.779 "io_queue_requests": 0, 00:24:47.779 "delay_cmd_submit": true, 00:24:47.779 "transport_retry_count": 4, 00:24:47.779 "bdev_retry_count": 3, 00:24:47.779 "transport_ack_timeout": 0, 00:24:47.779 "ctrlr_loss_timeout_sec": 0, 00:24:47.779 "reconnect_delay_sec": 0, 00:24:47.779 "fast_io_fail_timeout_sec": 0, 00:24:47.779 "disable_auto_failback": false, 00:24:47.779 "generate_uuids": false, 00:24:47.779 "transport_tos": 0, 00:24:47.779 "nvme_error_stat": false, 00:24:47.779 "rdma_srq_size": 0, 00:24:47.779 "io_path_stat": false, 00:24:47.779 "allow_accel_sequence": false, 00:24:47.779 "rdma_max_cq_size": 0, 00:24:47.779 "rdma_cm_event_timeout_ms": 0, 00:24:47.779 "dhchap_digests": [ 00:24:47.779 "sha256", 00:24:47.779 "sha384", 00:24:47.779 "sha512" 00:24:47.779 ], 00:24:47.779 "dhchap_dhgroups": [ 00:24:47.779 "null", 00:24:47.779 "ffdhe2048", 00:24:47.779 "ffdhe3072", 00:24:47.779 "ffdhe4096", 00:24:47.779 "ffdhe6144", 00:24:47.779 "ffdhe8192" 00:24:47.779 ] 00:24:47.779 } 00:24:47.779 }, 00:24:47.779 { 00:24:47.779 "method": "bdev_nvme_set_hotplug", 00:24:47.779 "params": { 00:24:47.779 "period_us": 100000, 00:24:47.779 "enable": false 00:24:47.779 } 00:24:47.779 }, 00:24:47.779 { 00:24:47.779 "method": "bdev_wait_for_examine" 00:24:47.779 } 00:24:47.779 ] 00:24:47.779 }, 00:24:47.779 { 00:24:47.779 "subsystem": "scsi", 00:24:47.779 "config": null 00:24:47.779 }, 00:24:47.779 { 00:24:47.779 "subsystem": "scheduler", 00:24:47.779 "config": [ 00:24:47.779 { 00:24:47.779 "method": "framework_set_scheduler", 00:24:47.779 "params": { 00:24:47.779 "name": "static" 00:24:47.779 } 00:24:47.779 } 00:24:47.779 ] 00:24:47.779 }, 00:24:47.779 { 00:24:47.779 "subsystem": "vhost_scsi", 00:24:47.779 "config": [] 00:24:47.779 }, 00:24:47.779 { 00:24:47.779 "subsystem": "vhost_blk", 00:24:47.779 "config": [] 00:24:47.779 }, 00:24:47.779 { 00:24:47.779 "subsystem": "ublk", 00:24:47.779 "config": [] 00:24:47.779 }, 00:24:47.779 { 00:24:47.779 "subsystem": "nbd", 00:24:47.779 "config": [] 00:24:47.779 }, 00:24:47.779 { 00:24:47.779 "subsystem": "nvmf", 00:24:47.779 "config": [ 00:24:47.779 { 00:24:47.779 "method": "nvmf_set_config", 00:24:47.779 "params": { 00:24:47.779 "discovery_filter": "match_any", 00:24:47.779 "admin_cmd_passthru": { 00:24:47.779 "identify_ctrlr": false 00:24:47.779 }, 00:24:47.779 "dhchap_digests": [ 00:24:47.779 "sha256", 00:24:47.779 "sha384", 00:24:47.779 "sha512" 00:24:47.779 ], 00:24:47.779 "dhchap_dhgroups": [ 00:24:47.779 "null", 00:24:47.779 "ffdhe2048", 00:24:47.779 "ffdhe3072", 00:24:47.779 "ffdhe4096", 00:24:47.779 "ffdhe6144", 00:24:47.779 "ffdhe8192" 00:24:47.779 ] 00:24:47.779 } 00:24:47.779 }, 00:24:47.779 { 00:24:47.779 "method": "nvmf_set_max_subsystems", 00:24:47.779 "params": { 00:24:47.779 "max_subsystems": 1024 00:24:47.779 } 00:24:47.779 }, 00:24:47.779 { 00:24:47.779 "method": "nvmf_set_crdt", 00:24:47.779 "params": { 00:24:47.779 "crdt1": 0, 00:24:47.779 "crdt2": 0, 00:24:47.779 "crdt3": 0 00:24:47.779 } 00:24:47.779 }, 00:24:47.779 { 00:24:47.779 "method": "nvmf_create_transport", 00:24:47.779 "params": { 00:24:47.779 "trtype": "TCP", 00:24:47.779 "max_queue_depth": 128, 00:24:47.779 "max_io_qpairs_per_ctrlr": 127, 00:24:47.779 "in_capsule_data_size": 4096, 00:24:47.779 "max_io_size": 131072, 00:24:47.779 "io_unit_size": 131072, 00:24:47.779 "max_aq_depth": 128, 00:24:47.779 "num_shared_buffers": 511, 00:24:47.779 "buf_cache_size": 4294967295, 00:24:47.779 "dif_insert_or_strip": false, 00:24:47.779 "zcopy": false, 00:24:47.779 "c2h_success": true, 00:24:47.779 "sock_priority": 0, 00:24:47.779 "abort_timeout_sec": 1, 00:24:47.779 "ack_timeout": 0, 00:24:47.779 "data_wr_pool_size": 0 00:24:47.779 } 00:24:47.779 } 00:24:47.779 ] 00:24:47.779 }, 00:24:47.779 { 00:24:47.779 "subsystem": "iscsi", 00:24:47.779 "config": [ 00:24:47.779 { 00:24:47.779 "method": "iscsi_set_options", 00:24:47.779 "params": { 00:24:47.779 "node_base": "iqn.2016-06.io.spdk", 00:24:47.779 "max_sessions": 128, 00:24:47.779 "max_connections_per_session": 2, 00:24:47.779 "max_queue_depth": 64, 00:24:47.779 "default_time2wait": 2, 00:24:47.779 "default_time2retain": 20, 00:24:47.779 "first_burst_length": 8192, 00:24:47.779 "immediate_data": true, 00:24:47.779 "allow_duplicated_isid": false, 00:24:47.779 "error_recovery_level": 0, 00:24:47.779 "nop_timeout": 60, 00:24:47.779 "nop_in_interval": 30, 00:24:47.779 "disable_chap": false, 00:24:47.779 "require_chap": false, 00:24:47.779 "mutual_chap": false, 00:24:47.779 "chap_group": 0, 00:24:47.779 "max_large_datain_per_connection": 64, 00:24:47.779 "max_r2t_per_connection": 4, 00:24:47.779 "pdu_pool_size": 36864, 00:24:47.779 "immediate_data_pool_size": 16384, 00:24:47.779 "data_out_pool_size": 2048 00:24:47.779 } 00:24:47.779 } 00:24:47.779 ] 00:24:47.779 } 00:24:47.779 ] 00:24:47.779 } 00:24:47.779 13:44:54 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@39 -- # trap - SIGINT SIGTERM EXIT 00:24:47.779 13:44:54 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@40 -- # killprocess 56414 00:24:47.779 13:44:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # '[' -z 56414 ']' 00:24:47.779 13:44:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@958 -- # kill -0 56414 00:24:47.779 13:44:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # uname 00:24:47.779 13:44:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:24:47.779 13:44:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 56414 00:24:47.779 13:44:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:24:47.779 killing process with pid 56414 00:24:47.779 13:44:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:24:47.779 13:44:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@972 -- # echo 'killing process with pid 56414' 00:24:47.779 13:44:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@973 -- # kill 56414 00:24:47.779 13:44:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@978 -- # wait 56414 00:24:49.153 13:44:55 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@47 -- # local spdk_pid=56454 00:24:49.153 13:44:55 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@46 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --json /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:24:49.153 13:44:55 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@48 -- # sleep 5 00:24:54.416 13:45:00 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@50 -- # killprocess 56454 00:24:54.416 13:45:00 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # '[' -z 56454 ']' 00:24:54.416 13:45:00 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@958 -- # kill -0 56454 00:24:54.416 13:45:00 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # uname 00:24:54.416 13:45:00 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:24:54.416 13:45:00 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 56454 00:24:54.416 13:45:00 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:24:54.416 killing process with pid 56454 00:24:54.416 13:45:00 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:24:54.416 13:45:00 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@972 -- # echo 'killing process with pid 56454' 00:24:54.416 13:45:00 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@973 -- # kill 56454 00:24:54.416 13:45:00 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@978 -- # wait 56454 00:24:55.789 13:45:02 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@51 -- # grep -q 'TCP Transport Init' /home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:24:55.789 13:45:02 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@52 -- # rm /home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:24:55.789 00:24:55.789 real 0m9.003s 00:24:55.789 user 0m8.540s 00:24:55.789 sys 0m0.639s 00:24:55.789 13:45:02 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1130 -- # xtrace_disable 00:24:55.789 13:45:02 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:24:55.789 ************************************ 00:24:55.789 END TEST skip_rpc_with_json 00:24:55.789 ************************************ 00:24:55.789 13:45:02 skip_rpc -- rpc/skip_rpc.sh@75 -- # run_test skip_rpc_with_delay test_skip_rpc_with_delay 00:24:55.789 13:45:02 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:24:55.789 13:45:02 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:24:55.789 13:45:02 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:24:55.789 ************************************ 00:24:55.789 START TEST skip_rpc_with_delay 00:24:55.789 ************************************ 00:24:55.789 13:45:02 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1129 -- # test_skip_rpc_with_delay 00:24:55.790 13:45:02 skip_rpc.skip_rpc_with_delay -- rpc/skip_rpc.sh@57 -- # NOT /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:24:55.790 13:45:02 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@652 -- # local es=0 00:24:55.790 13:45:02 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@654 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:24:55.790 13:45:02 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@640 -- # local arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:24:55.790 13:45:02 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:24:55.790 13:45:02 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # type -t /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:24:55.790 13:45:02 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:24:55.790 13:45:02 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # type -P /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:24:55.790 13:45:02 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:24:55.790 13:45:02 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:24:55.790 13:45:02 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # [[ -x /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt ]] 00:24:55.790 13:45:02 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@655 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:24:55.790 [2024-11-20 13:45:02.648284] app.c: 842:spdk_app_start: *ERROR*: Cannot use '--wait-for-rpc' if no RPC server is going to be started. 00:24:55.790 13:45:02 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@655 -- # es=1 00:24:55.790 13:45:02 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:24:55.790 13:45:02 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:24:55.790 13:45:02 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:24:55.790 00:24:55.790 real 0m0.125s 00:24:55.790 user 0m0.067s 00:24:55.790 sys 0m0.057s 00:24:55.790 ************************************ 00:24:55.790 END TEST skip_rpc_with_delay 00:24:55.790 13:45:02 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1130 -- # xtrace_disable 00:24:55.790 13:45:02 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@10 -- # set +x 00:24:55.790 ************************************ 00:24:55.790 13:45:02 skip_rpc -- rpc/skip_rpc.sh@77 -- # uname 00:24:55.790 13:45:02 skip_rpc -- rpc/skip_rpc.sh@77 -- # '[' Linux '!=' FreeBSD ']' 00:24:55.790 13:45:02 skip_rpc -- rpc/skip_rpc.sh@78 -- # run_test exit_on_failed_rpc_init test_exit_on_failed_rpc_init 00:24:55.790 13:45:02 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:24:55.790 13:45:02 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:24:55.790 13:45:02 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:24:55.790 ************************************ 00:24:55.790 START TEST exit_on_failed_rpc_init 00:24:55.790 ************************************ 00:24:55.790 13:45:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1129 -- # test_exit_on_failed_rpc_init 00:24:55.790 13:45:02 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@62 -- # local spdk_pid=56576 00:24:55.790 13:45:02 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@63 -- # waitforlisten 56576 00:24:55.790 13:45:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@835 -- # '[' -z 56576 ']' 00:24:55.790 13:45:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:55.790 13:45:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@840 -- # local max_retries=100 00:24:55.790 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:55.790 13:45:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:55.790 13:45:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@844 -- # xtrace_disable 00:24:55.790 13:45:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:24:55.790 13:45:02 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@61 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:24:55.790 [2024-11-20 13:45:02.815300] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:24:55.790 [2024-11-20 13:45:02.815440] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid56576 ] 00:24:56.048 [2024-11-20 13:45:02.977973] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:56.048 [2024-11-20 13:45:03.098193] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:24:56.982 13:45:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:24:56.982 13:45:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@868 -- # return 0 00:24:56.982 13:45:03 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@65 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:24:56.982 13:45:03 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@67 -- # NOT /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:24:56.982 13:45:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@652 -- # local es=0 00:24:56.982 13:45:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@654 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:24:56.982 13:45:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@640 -- # local arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:24:56.982 13:45:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:24:56.982 13:45:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # type -t /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:24:56.982 13:45:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:24:56.982 13:45:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # type -P /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:24:56.982 13:45:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:24:56.982 13:45:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:24:56.982 13:45:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # [[ -x /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt ]] 00:24:56.982 13:45:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@655 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:24:56.982 [2024-11-20 13:45:03.830199] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:24:56.982 [2024-11-20 13:45:03.830314] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid56594 ] 00:24:56.982 [2024-11-20 13:45:03.989789] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:57.239 [2024-11-20 13:45:04.107409] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:24:57.239 [2024-11-20 13:45:04.107514] rpc.c: 180:_spdk_rpc_listen: *ERROR*: RPC Unix domain socket path /var/tmp/spdk.sock in use. Specify another. 00:24:57.239 [2024-11-20 13:45:04.107529] rpc.c: 166:spdk_rpc_initialize: *ERROR*: Unable to start RPC service at /var/tmp/spdk.sock 00:24:57.239 [2024-11-20 13:45:04.107541] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:24:57.498 13:45:04 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@655 -- # es=234 00:24:57.498 13:45:04 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:24:57.498 13:45:04 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@664 -- # es=106 00:24:57.498 13:45:04 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@665 -- # case "$es" in 00:24:57.498 13:45:04 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@672 -- # es=1 00:24:57.498 13:45:04 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:24:57.498 13:45:04 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:24:57.498 13:45:04 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@70 -- # killprocess 56576 00:24:57.498 13:45:04 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@954 -- # '[' -z 56576 ']' 00:24:57.498 13:45:04 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@958 -- # kill -0 56576 00:24:57.498 13:45:04 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@959 -- # uname 00:24:57.498 13:45:04 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:24:57.498 13:45:04 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 56576 00:24:57.498 killing process with pid 56576 00:24:57.498 13:45:04 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:24:57.498 13:45:04 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:24:57.498 13:45:04 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@972 -- # echo 'killing process with pid 56576' 00:24:57.498 13:45:04 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@973 -- # kill 56576 00:24:57.498 13:45:04 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@978 -- # wait 56576 00:24:58.870 00:24:58.870 real 0m3.186s 00:24:58.870 user 0m3.472s 00:24:58.870 sys 0m0.446s 00:24:58.870 13:45:05 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1130 -- # xtrace_disable 00:24:58.870 ************************************ 00:24:58.870 END TEST exit_on_failed_rpc_init 00:24:58.870 ************************************ 00:24:58.870 13:45:05 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:24:59.128 13:45:05 skip_rpc -- rpc/skip_rpc.sh@81 -- # rm /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:24:59.128 00:24:59.128 real 0m18.946s 00:24:59.128 user 0m18.092s 00:24:59.128 sys 0m1.656s 00:24:59.128 13:45:05 skip_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:24:59.128 13:45:05 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:24:59.128 ************************************ 00:24:59.128 END TEST skip_rpc 00:24:59.128 ************************************ 00:24:59.128 13:45:05 -- spdk/autotest.sh@158 -- # run_test rpc_client /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client.sh 00:24:59.128 13:45:05 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:24:59.128 13:45:05 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:24:59.128 13:45:05 -- common/autotest_common.sh@10 -- # set +x 00:24:59.128 ************************************ 00:24:59.128 START TEST rpc_client 00:24:59.128 ************************************ 00:24:59.128 13:45:05 rpc_client -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client.sh 00:24:59.128 * Looking for test storage... 00:24:59.128 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc_client 00:24:59.128 13:45:06 rpc_client -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:24:59.128 13:45:06 rpc_client -- common/autotest_common.sh@1693 -- # lcov --version 00:24:59.128 13:45:06 rpc_client -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:24:59.128 13:45:06 rpc_client -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@333 -- # local ver1 ver1_l 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@334 -- # local ver2 ver2_l 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@336 -- # IFS=.-: 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@336 -- # read -ra ver1 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@337 -- # IFS=.-: 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@337 -- # read -ra ver2 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@338 -- # local 'op=<' 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@340 -- # ver1_l=2 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@341 -- # ver2_l=1 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@344 -- # case "$op" in 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@345 -- # : 1 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@364 -- # (( v = 0 )) 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@365 -- # decimal 1 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@353 -- # local d=1 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@355 -- # echo 1 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@365 -- # ver1[v]=1 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@366 -- # decimal 2 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@353 -- # local d=2 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@355 -- # echo 2 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@366 -- # ver2[v]=2 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:24:59.128 13:45:06 rpc_client -- scripts/common.sh@368 -- # return 0 00:24:59.128 13:45:06 rpc_client -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:24:59.128 13:45:06 rpc_client -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:24:59.128 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:59.128 --rc genhtml_branch_coverage=1 00:24:59.128 --rc genhtml_function_coverage=1 00:24:59.128 --rc genhtml_legend=1 00:24:59.128 --rc geninfo_all_blocks=1 00:24:59.128 --rc geninfo_unexecuted_blocks=1 00:24:59.128 00:24:59.128 ' 00:24:59.128 13:45:06 rpc_client -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:24:59.128 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:59.128 --rc genhtml_branch_coverage=1 00:24:59.128 --rc genhtml_function_coverage=1 00:24:59.128 --rc genhtml_legend=1 00:24:59.128 --rc geninfo_all_blocks=1 00:24:59.128 --rc geninfo_unexecuted_blocks=1 00:24:59.128 00:24:59.128 ' 00:24:59.128 13:45:06 rpc_client -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:24:59.128 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:59.128 --rc genhtml_branch_coverage=1 00:24:59.128 --rc genhtml_function_coverage=1 00:24:59.128 --rc genhtml_legend=1 00:24:59.128 --rc geninfo_all_blocks=1 00:24:59.128 --rc geninfo_unexecuted_blocks=1 00:24:59.128 00:24:59.128 ' 00:24:59.128 13:45:06 rpc_client -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:24:59.128 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:59.128 --rc genhtml_branch_coverage=1 00:24:59.128 --rc genhtml_function_coverage=1 00:24:59.128 --rc genhtml_legend=1 00:24:59.128 --rc geninfo_all_blocks=1 00:24:59.128 --rc geninfo_unexecuted_blocks=1 00:24:59.128 00:24:59.128 ' 00:24:59.129 13:45:06 rpc_client -- rpc_client/rpc_client.sh@10 -- # /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client_test 00:24:59.129 OK 00:24:59.387 13:45:06 rpc_client -- rpc_client/rpc_client.sh@12 -- # trap - SIGINT SIGTERM EXIT 00:24:59.387 00:24:59.387 real 0m0.197s 00:24:59.387 user 0m0.120s 00:24:59.387 sys 0m0.087s 00:24:59.387 13:45:06 rpc_client -- common/autotest_common.sh@1130 -- # xtrace_disable 00:24:59.387 13:45:06 rpc_client -- common/autotest_common.sh@10 -- # set +x 00:24:59.387 ************************************ 00:24:59.387 END TEST rpc_client 00:24:59.387 ************************************ 00:24:59.387 13:45:06 -- spdk/autotest.sh@159 -- # run_test json_config /home/vagrant/spdk_repo/spdk/test/json_config/json_config.sh 00:24:59.387 13:45:06 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:24:59.387 13:45:06 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:24:59.387 13:45:06 -- common/autotest_common.sh@10 -- # set +x 00:24:59.387 ************************************ 00:24:59.387 START TEST json_config 00:24:59.387 ************************************ 00:24:59.387 13:45:06 json_config -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/json_config/json_config.sh 00:24:59.387 13:45:06 json_config -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:24:59.387 13:45:06 json_config -- common/autotest_common.sh@1693 -- # lcov --version 00:24:59.387 13:45:06 json_config -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:24:59.387 13:45:06 json_config -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:24:59.387 13:45:06 json_config -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:24:59.387 13:45:06 json_config -- scripts/common.sh@333 -- # local ver1 ver1_l 00:24:59.387 13:45:06 json_config -- scripts/common.sh@334 -- # local ver2 ver2_l 00:24:59.387 13:45:06 json_config -- scripts/common.sh@336 -- # IFS=.-: 00:24:59.387 13:45:06 json_config -- scripts/common.sh@336 -- # read -ra ver1 00:24:59.387 13:45:06 json_config -- scripts/common.sh@337 -- # IFS=.-: 00:24:59.387 13:45:06 json_config -- scripts/common.sh@337 -- # read -ra ver2 00:24:59.387 13:45:06 json_config -- scripts/common.sh@338 -- # local 'op=<' 00:24:59.387 13:45:06 json_config -- scripts/common.sh@340 -- # ver1_l=2 00:24:59.387 13:45:06 json_config -- scripts/common.sh@341 -- # ver2_l=1 00:24:59.387 13:45:06 json_config -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:24:59.387 13:45:06 json_config -- scripts/common.sh@344 -- # case "$op" in 00:24:59.387 13:45:06 json_config -- scripts/common.sh@345 -- # : 1 00:24:59.387 13:45:06 json_config -- scripts/common.sh@364 -- # (( v = 0 )) 00:24:59.387 13:45:06 json_config -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:24:59.387 13:45:06 json_config -- scripts/common.sh@365 -- # decimal 1 00:24:59.387 13:45:06 json_config -- scripts/common.sh@353 -- # local d=1 00:24:59.387 13:45:06 json_config -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:24:59.387 13:45:06 json_config -- scripts/common.sh@355 -- # echo 1 00:24:59.387 13:45:06 json_config -- scripts/common.sh@365 -- # ver1[v]=1 00:24:59.387 13:45:06 json_config -- scripts/common.sh@366 -- # decimal 2 00:24:59.387 13:45:06 json_config -- scripts/common.sh@353 -- # local d=2 00:24:59.387 13:45:06 json_config -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:24:59.387 13:45:06 json_config -- scripts/common.sh@355 -- # echo 2 00:24:59.387 13:45:06 json_config -- scripts/common.sh@366 -- # ver2[v]=2 00:24:59.387 13:45:06 json_config -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:24:59.387 13:45:06 json_config -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:24:59.387 13:45:06 json_config -- scripts/common.sh@368 -- # return 0 00:24:59.387 13:45:06 json_config -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:24:59.387 13:45:06 json_config -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:24:59.387 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:59.387 --rc genhtml_branch_coverage=1 00:24:59.387 --rc genhtml_function_coverage=1 00:24:59.387 --rc genhtml_legend=1 00:24:59.387 --rc geninfo_all_blocks=1 00:24:59.387 --rc geninfo_unexecuted_blocks=1 00:24:59.387 00:24:59.387 ' 00:24:59.387 13:45:06 json_config -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:24:59.387 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:59.387 --rc genhtml_branch_coverage=1 00:24:59.387 --rc genhtml_function_coverage=1 00:24:59.387 --rc genhtml_legend=1 00:24:59.387 --rc geninfo_all_blocks=1 00:24:59.387 --rc geninfo_unexecuted_blocks=1 00:24:59.387 00:24:59.387 ' 00:24:59.387 13:45:06 json_config -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:24:59.387 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:59.387 --rc genhtml_branch_coverage=1 00:24:59.387 --rc genhtml_function_coverage=1 00:24:59.387 --rc genhtml_legend=1 00:24:59.387 --rc geninfo_all_blocks=1 00:24:59.387 --rc geninfo_unexecuted_blocks=1 00:24:59.387 00:24:59.387 ' 00:24:59.387 13:45:06 json_config -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:24:59.387 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:59.387 --rc genhtml_branch_coverage=1 00:24:59.387 --rc genhtml_function_coverage=1 00:24:59.387 --rc genhtml_legend=1 00:24:59.387 --rc geninfo_all_blocks=1 00:24:59.387 --rc geninfo_unexecuted_blocks=1 00:24:59.387 00:24:59.387 ' 00:24:59.387 13:45:06 json_config -- json_config/json_config.sh@8 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:24:59.387 13:45:06 json_config -- nvmf/common.sh@7 -- # uname -s 00:24:59.387 13:45:06 json_config -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:24:59.387 13:45:06 json_config -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:24:59.387 13:45:06 json_config -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:24:59.387 13:45:06 json_config -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:24:59.387 13:45:06 json_config -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:24:59.387 13:45:06 json_config -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:24:59.387 13:45:06 json_config -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:24:59.387 13:45:06 json_config -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:24:59.387 13:45:06 json_config -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:24:59.387 13:45:06 json_config -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:24:59.387 13:45:06 json_config -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:0780082f-c229-467d-8bb4-b14a621a20b9 00:24:59.387 13:45:06 json_config -- nvmf/common.sh@18 -- # NVME_HOSTID=0780082f-c229-467d-8bb4-b14a621a20b9 00:24:59.387 13:45:06 json_config -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:24:59.387 13:45:06 json_config -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:24:59.387 13:45:06 json_config -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:24:59.387 13:45:06 json_config -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:24:59.387 13:45:06 json_config -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:24:59.387 13:45:06 json_config -- scripts/common.sh@15 -- # shopt -s extglob 00:24:59.387 13:45:06 json_config -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:24:59.387 13:45:06 json_config -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:24:59.387 13:45:06 json_config -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:24:59.387 13:45:06 json_config -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:59.388 13:45:06 json_config -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:59.388 13:45:06 json_config -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:59.388 13:45:06 json_config -- paths/export.sh@5 -- # export PATH 00:24:59.388 13:45:06 json_config -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:59.388 13:45:06 json_config -- nvmf/common.sh@51 -- # : 0 00:24:59.388 13:45:06 json_config -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:24:59.388 13:45:06 json_config -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:24:59.388 13:45:06 json_config -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:24:59.388 13:45:06 json_config -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:24:59.388 13:45:06 json_config -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:24:59.388 13:45:06 json_config -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:24:59.388 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:24:59.388 13:45:06 json_config -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:24:59.388 13:45:06 json_config -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:24:59.388 13:45:06 json_config -- nvmf/common.sh@55 -- # have_pci_nics=0 00:24:59.388 13:45:06 json_config -- json_config/json_config.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/json_config/common.sh 00:24:59.388 13:45:06 json_config -- json_config/json_config.sh@11 -- # [[ 0 -eq 1 ]] 00:24:59.388 13:45:06 json_config -- json_config/json_config.sh@15 -- # [[ 0 -ne 1 ]] 00:24:59.388 13:45:06 json_config -- json_config/json_config.sh@15 -- # [[ 0 -eq 1 ]] 00:24:59.388 13:45:06 json_config -- json_config/json_config.sh@26 -- # (( SPDK_TEST_BLOCKDEV + SPDK_TEST_ISCSI + SPDK_TEST_NVMF + SPDK_TEST_VHOST + SPDK_TEST_VHOST_INIT + SPDK_TEST_RBD == 0 )) 00:24:59.388 13:45:06 json_config -- json_config/json_config.sh@27 -- # echo 'WARNING: No tests are enabled so not running JSON configuration tests' 00:24:59.388 WARNING: No tests are enabled so not running JSON configuration tests 00:24:59.388 13:45:06 json_config -- json_config/json_config.sh@28 -- # exit 0 00:24:59.388 00:24:59.388 real 0m0.145s 00:24:59.388 user 0m0.088s 00:24:59.388 sys 0m0.059s 00:24:59.388 13:45:06 json_config -- common/autotest_common.sh@1130 -- # xtrace_disable 00:24:59.388 13:45:06 json_config -- common/autotest_common.sh@10 -- # set +x 00:24:59.388 ************************************ 00:24:59.388 END TEST json_config 00:24:59.388 ************************************ 00:24:59.388 13:45:06 -- spdk/autotest.sh@160 -- # run_test json_config_extra_key /home/vagrant/spdk_repo/spdk/test/json_config/json_config_extra_key.sh 00:24:59.388 13:45:06 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:24:59.388 13:45:06 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:24:59.388 13:45:06 -- common/autotest_common.sh@10 -- # set +x 00:24:59.388 ************************************ 00:24:59.388 START TEST json_config_extra_key 00:24:59.388 ************************************ 00:24:59.388 13:45:06 json_config_extra_key -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/json_config/json_config_extra_key.sh 00:24:59.646 13:45:06 json_config_extra_key -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:24:59.646 13:45:06 json_config_extra_key -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:24:59.646 13:45:06 json_config_extra_key -- common/autotest_common.sh@1693 -- # lcov --version 00:24:59.646 13:45:06 json_config_extra_key -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@333 -- # local ver1 ver1_l 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@334 -- # local ver2 ver2_l 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@336 -- # IFS=.-: 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@336 -- # read -ra ver1 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@337 -- # IFS=.-: 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@337 -- # read -ra ver2 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@338 -- # local 'op=<' 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@340 -- # ver1_l=2 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@341 -- # ver2_l=1 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@344 -- # case "$op" in 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@345 -- # : 1 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@364 -- # (( v = 0 )) 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@365 -- # decimal 1 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@353 -- # local d=1 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@355 -- # echo 1 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@365 -- # ver1[v]=1 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@366 -- # decimal 2 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@353 -- # local d=2 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@355 -- # echo 2 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@366 -- # ver2[v]=2 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:24:59.646 13:45:06 json_config_extra_key -- scripts/common.sh@368 -- # return 0 00:24:59.646 13:45:06 json_config_extra_key -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:24:59.646 13:45:06 json_config_extra_key -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:24:59.646 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:59.646 --rc genhtml_branch_coverage=1 00:24:59.646 --rc genhtml_function_coverage=1 00:24:59.646 --rc genhtml_legend=1 00:24:59.646 --rc geninfo_all_blocks=1 00:24:59.646 --rc geninfo_unexecuted_blocks=1 00:24:59.646 00:24:59.646 ' 00:24:59.646 13:45:06 json_config_extra_key -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:24:59.646 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:59.646 --rc genhtml_branch_coverage=1 00:24:59.646 --rc genhtml_function_coverage=1 00:24:59.646 --rc genhtml_legend=1 00:24:59.646 --rc geninfo_all_blocks=1 00:24:59.646 --rc geninfo_unexecuted_blocks=1 00:24:59.646 00:24:59.646 ' 00:24:59.646 13:45:06 json_config_extra_key -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:24:59.646 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:59.646 --rc genhtml_branch_coverage=1 00:24:59.646 --rc genhtml_function_coverage=1 00:24:59.646 --rc genhtml_legend=1 00:24:59.646 --rc geninfo_all_blocks=1 00:24:59.646 --rc geninfo_unexecuted_blocks=1 00:24:59.646 00:24:59.646 ' 00:24:59.646 13:45:06 json_config_extra_key -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:24:59.646 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:59.647 --rc genhtml_branch_coverage=1 00:24:59.647 --rc genhtml_function_coverage=1 00:24:59.647 --rc genhtml_legend=1 00:24:59.647 --rc geninfo_all_blocks=1 00:24:59.647 --rc geninfo_unexecuted_blocks=1 00:24:59.647 00:24:59.647 ' 00:24:59.647 13:45:06 json_config_extra_key -- json_config/json_config_extra_key.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@7 -- # uname -s 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:0780082f-c229-467d-8bb4-b14a621a20b9 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@18 -- # NVME_HOSTID=0780082f-c229-467d-8bb4-b14a621a20b9 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:24:59.647 13:45:06 json_config_extra_key -- scripts/common.sh@15 -- # shopt -s extglob 00:24:59.647 13:45:06 json_config_extra_key -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:24:59.647 13:45:06 json_config_extra_key -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:24:59.647 13:45:06 json_config_extra_key -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:24:59.647 13:45:06 json_config_extra_key -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:59.647 13:45:06 json_config_extra_key -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:59.647 13:45:06 json_config_extra_key -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:59.647 13:45:06 json_config_extra_key -- paths/export.sh@5 -- # export PATH 00:24:59.647 13:45:06 json_config_extra_key -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@51 -- # : 0 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:24:59.647 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:24:59.647 13:45:06 json_config_extra_key -- nvmf/common.sh@55 -- # have_pci_nics=0 00:24:59.647 13:45:06 json_config_extra_key -- json_config/json_config_extra_key.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/json_config/common.sh 00:24:59.647 13:45:06 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # app_pid=(['target']='') 00:24:59.647 13:45:06 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # declare -A app_pid 00:24:59.647 13:45:06 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # app_socket=(['target']='/var/tmp/spdk_tgt.sock') 00:24:59.647 13:45:06 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # declare -A app_socket 00:24:59.647 13:45:06 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # app_params=(['target']='-m 0x1 -s 1024') 00:24:59.647 13:45:06 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # declare -A app_params 00:24:59.647 13:45:06 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # configs_path=(['target']='/home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json') 00:24:59.647 13:45:06 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # declare -A configs_path 00:24:59.647 13:45:06 json_config_extra_key -- json_config/json_config_extra_key.sh@22 -- # trap 'on_error_exit "${FUNCNAME}" "${LINENO}"' ERR 00:24:59.647 13:45:06 json_config_extra_key -- json_config/json_config_extra_key.sh@24 -- # echo 'INFO: launching applications...' 00:24:59.647 INFO: launching applications... 00:24:59.647 13:45:06 json_config_extra_key -- json_config/json_config_extra_key.sh@25 -- # json_config_test_start_app target --json /home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json 00:24:59.647 13:45:06 json_config_extra_key -- json_config/common.sh@9 -- # local app=target 00:24:59.647 13:45:06 json_config_extra_key -- json_config/common.sh@10 -- # shift 00:24:59.647 13:45:06 json_config_extra_key -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:24:59.647 13:45:06 json_config_extra_key -- json_config/common.sh@13 -- # [[ -z '' ]] 00:24:59.647 13:45:06 json_config_extra_key -- json_config/common.sh@15 -- # local app_extra_params= 00:24:59.647 13:45:06 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:24:59.647 13:45:06 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:24:59.647 13:45:06 json_config_extra_key -- json_config/common.sh@22 -- # app_pid["$app"]=56793 00:24:59.647 13:45:06 json_config_extra_key -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:24:59.647 Waiting for target to run... 00:24:59.647 13:45:06 json_config_extra_key -- json_config/common.sh@25 -- # waitforlisten 56793 /var/tmp/spdk_tgt.sock 00:24:59.647 13:45:06 json_config_extra_key -- common/autotest_common.sh@835 -- # '[' -z 56793 ']' 00:24:59.647 13:45:06 json_config_extra_key -- json_config/common.sh@21 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --json /home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json 00:24:59.647 13:45:06 json_config_extra_key -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:24:59.647 13:45:06 json_config_extra_key -- common/autotest_common.sh@840 -- # local max_retries=100 00:24:59.647 13:45:06 json_config_extra_key -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:24:59.647 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:24:59.647 13:45:06 json_config_extra_key -- common/autotest_common.sh@844 -- # xtrace_disable 00:24:59.647 13:45:06 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:24:59.647 [2024-11-20 13:45:06.650535] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:24:59.647 [2024-11-20 13:45:06.650795] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid56793 ] 00:25:00.212 [2024-11-20 13:45:07.030498] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:00.212 [2024-11-20 13:45:07.126312] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:00.778 13:45:07 json_config_extra_key -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:25:00.778 13:45:07 json_config_extra_key -- common/autotest_common.sh@868 -- # return 0 00:25:00.778 13:45:07 json_config_extra_key -- json_config/common.sh@26 -- # echo '' 00:25:00.778 00:25:00.778 INFO: shutting down applications... 00:25:00.778 13:45:07 json_config_extra_key -- json_config/json_config_extra_key.sh@27 -- # echo 'INFO: shutting down applications...' 00:25:00.778 13:45:07 json_config_extra_key -- json_config/json_config_extra_key.sh@28 -- # json_config_test_shutdown_app target 00:25:00.778 13:45:07 json_config_extra_key -- json_config/common.sh@31 -- # local app=target 00:25:00.778 13:45:07 json_config_extra_key -- json_config/common.sh@34 -- # [[ -n 22 ]] 00:25:00.778 13:45:07 json_config_extra_key -- json_config/common.sh@35 -- # [[ -n 56793 ]] 00:25:00.778 13:45:07 json_config_extra_key -- json_config/common.sh@38 -- # kill -SIGINT 56793 00:25:00.778 13:45:07 json_config_extra_key -- json_config/common.sh@40 -- # (( i = 0 )) 00:25:00.778 13:45:07 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:25:00.778 13:45:07 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 56793 00:25:00.778 13:45:07 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:25:01.344 13:45:08 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:25:01.344 13:45:08 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:25:01.344 13:45:08 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 56793 00:25:01.344 13:45:08 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:25:01.910 13:45:08 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:25:01.910 13:45:08 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:25:01.910 13:45:08 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 56793 00:25:01.910 13:45:08 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:25:02.168 13:45:09 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:25:02.168 13:45:09 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:25:02.168 13:45:09 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 56793 00:25:02.168 13:45:09 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:25:02.734 13:45:09 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:25:02.734 13:45:09 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:25:02.734 13:45:09 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 56793 00:25:02.734 13:45:09 json_config_extra_key -- json_config/common.sh@42 -- # app_pid["$app"]= 00:25:02.734 13:45:09 json_config_extra_key -- json_config/common.sh@43 -- # break 00:25:02.734 SPDK target shutdown done 00:25:02.734 13:45:09 json_config_extra_key -- json_config/common.sh@48 -- # [[ -n '' ]] 00:25:02.734 13:45:09 json_config_extra_key -- json_config/common.sh@53 -- # echo 'SPDK target shutdown done' 00:25:02.734 Success 00:25:02.734 13:45:09 json_config_extra_key -- json_config/json_config_extra_key.sh@30 -- # echo Success 00:25:02.734 ************************************ 00:25:02.734 END TEST json_config_extra_key 00:25:02.734 ************************************ 00:25:02.734 00:25:02.734 real 0m3.264s 00:25:02.734 user 0m2.988s 00:25:02.734 sys 0m0.465s 00:25:02.734 13:45:09 json_config_extra_key -- common/autotest_common.sh@1130 -- # xtrace_disable 00:25:02.734 13:45:09 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:25:02.734 13:45:09 -- spdk/autotest.sh@161 -- # run_test alias_rpc /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:25:02.734 13:45:09 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:25:02.734 13:45:09 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:25:02.734 13:45:09 -- common/autotest_common.sh@10 -- # set +x 00:25:02.734 ************************************ 00:25:02.734 START TEST alias_rpc 00:25:02.734 ************************************ 00:25:02.734 13:45:09 alias_rpc -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:25:02.992 * Looking for test storage... 00:25:02.992 * Found test storage at /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc 00:25:02.992 13:45:09 alias_rpc -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:25:02.992 13:45:09 alias_rpc -- common/autotest_common.sh@1693 -- # lcov --version 00:25:02.992 13:45:09 alias_rpc -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:25:02.992 13:45:09 alias_rpc -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@344 -- # case "$op" in 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@345 -- # : 1 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@365 -- # decimal 1 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@353 -- # local d=1 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@355 -- # echo 1 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@366 -- # decimal 2 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@353 -- # local d=2 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@355 -- # echo 2 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:25:02.992 13:45:09 alias_rpc -- scripts/common.sh@368 -- # return 0 00:25:02.992 13:45:09 alias_rpc -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:25:02.992 13:45:09 alias_rpc -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:25:02.992 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:02.992 --rc genhtml_branch_coverage=1 00:25:02.992 --rc genhtml_function_coverage=1 00:25:02.992 --rc genhtml_legend=1 00:25:02.992 --rc geninfo_all_blocks=1 00:25:02.992 --rc geninfo_unexecuted_blocks=1 00:25:02.992 00:25:02.992 ' 00:25:02.992 13:45:09 alias_rpc -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:25:02.992 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:02.992 --rc genhtml_branch_coverage=1 00:25:02.992 --rc genhtml_function_coverage=1 00:25:02.992 --rc genhtml_legend=1 00:25:02.992 --rc geninfo_all_blocks=1 00:25:02.992 --rc geninfo_unexecuted_blocks=1 00:25:02.992 00:25:02.992 ' 00:25:02.992 13:45:09 alias_rpc -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:25:02.992 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:02.992 --rc genhtml_branch_coverage=1 00:25:02.992 --rc genhtml_function_coverage=1 00:25:02.992 --rc genhtml_legend=1 00:25:02.992 --rc geninfo_all_blocks=1 00:25:02.992 --rc geninfo_unexecuted_blocks=1 00:25:02.992 00:25:02.992 ' 00:25:02.992 13:45:09 alias_rpc -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:25:02.992 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:02.992 --rc genhtml_branch_coverage=1 00:25:02.992 --rc genhtml_function_coverage=1 00:25:02.992 --rc genhtml_legend=1 00:25:02.992 --rc geninfo_all_blocks=1 00:25:02.992 --rc geninfo_unexecuted_blocks=1 00:25:02.992 00:25:02.992 ' 00:25:02.992 13:45:09 alias_rpc -- alias_rpc/alias_rpc.sh@10 -- # trap 'killprocess $spdk_tgt_pid; exit 1' ERR 00:25:02.992 13:45:09 alias_rpc -- alias_rpc/alias_rpc.sh@13 -- # spdk_tgt_pid=56886 00:25:02.992 13:45:09 alias_rpc -- alias_rpc/alias_rpc.sh@14 -- # waitforlisten 56886 00:25:02.992 13:45:09 alias_rpc -- alias_rpc/alias_rpc.sh@12 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:25:02.992 13:45:09 alias_rpc -- common/autotest_common.sh@835 -- # '[' -z 56886 ']' 00:25:02.992 13:45:09 alias_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:02.992 13:45:09 alias_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:25:02.992 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:02.992 13:45:09 alias_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:02.992 13:45:09 alias_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:25:02.992 13:45:09 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:25:02.992 [2024-11-20 13:45:09.966517] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:25:02.992 [2024-11-20 13:45:09.966781] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid56886 ] 00:25:03.250 [2024-11-20 13:45:10.130649] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:03.250 [2024-11-20 13:45:10.251901] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:04.183 13:45:10 alias_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:25:04.183 13:45:10 alias_rpc -- common/autotest_common.sh@868 -- # return 0 00:25:04.183 13:45:10 alias_rpc -- alias_rpc/alias_rpc.sh@17 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py load_config -i 00:25:04.183 13:45:11 alias_rpc -- alias_rpc/alias_rpc.sh@19 -- # killprocess 56886 00:25:04.183 13:45:11 alias_rpc -- common/autotest_common.sh@954 -- # '[' -z 56886 ']' 00:25:04.183 13:45:11 alias_rpc -- common/autotest_common.sh@958 -- # kill -0 56886 00:25:04.183 13:45:11 alias_rpc -- common/autotest_common.sh@959 -- # uname 00:25:04.183 13:45:11 alias_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:25:04.184 13:45:11 alias_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 56886 00:25:04.184 killing process with pid 56886 00:25:04.184 13:45:11 alias_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:25:04.184 13:45:11 alias_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:25:04.184 13:45:11 alias_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 56886' 00:25:04.184 13:45:11 alias_rpc -- common/autotest_common.sh@973 -- # kill 56886 00:25:04.184 13:45:11 alias_rpc -- common/autotest_common.sh@978 -- # wait 56886 00:25:06.183 ************************************ 00:25:06.183 END TEST alias_rpc 00:25:06.183 ************************************ 00:25:06.183 00:25:06.183 real 0m3.086s 00:25:06.183 user 0m3.121s 00:25:06.183 sys 0m0.491s 00:25:06.183 13:45:12 alias_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:25:06.183 13:45:12 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:25:06.183 13:45:12 -- spdk/autotest.sh@163 -- # [[ 0 -eq 0 ]] 00:25:06.183 13:45:12 -- spdk/autotest.sh@164 -- # run_test spdkcli_tcp /home/vagrant/spdk_repo/spdk/test/spdkcli/tcp.sh 00:25:06.183 13:45:12 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:25:06.183 13:45:12 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:25:06.183 13:45:12 -- common/autotest_common.sh@10 -- # set +x 00:25:06.183 ************************************ 00:25:06.183 START TEST spdkcli_tcp 00:25:06.183 ************************************ 00:25:06.183 13:45:12 spdkcli_tcp -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/tcp.sh 00:25:06.183 * Looking for test storage... 00:25:06.183 * Found test storage at /home/vagrant/spdk_repo/spdk/test/spdkcli 00:25:06.183 13:45:12 spdkcli_tcp -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:25:06.183 13:45:12 spdkcli_tcp -- common/autotest_common.sh@1693 -- # lcov --version 00:25:06.183 13:45:12 spdkcli_tcp -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:25:06.183 13:45:12 spdkcli_tcp -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:25:06.183 13:45:12 spdkcli_tcp -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:25:06.183 13:45:12 spdkcli_tcp -- scripts/common.sh@333 -- # local ver1 ver1_l 00:25:06.183 13:45:12 spdkcli_tcp -- scripts/common.sh@334 -- # local ver2 ver2_l 00:25:06.183 13:45:12 spdkcli_tcp -- scripts/common.sh@336 -- # IFS=.-: 00:25:06.183 13:45:12 spdkcli_tcp -- scripts/common.sh@336 -- # read -ra ver1 00:25:06.183 13:45:12 spdkcli_tcp -- scripts/common.sh@337 -- # IFS=.-: 00:25:06.183 13:45:12 spdkcli_tcp -- scripts/common.sh@337 -- # read -ra ver2 00:25:06.183 13:45:12 spdkcli_tcp -- scripts/common.sh@338 -- # local 'op=<' 00:25:06.183 13:45:13 spdkcli_tcp -- scripts/common.sh@340 -- # ver1_l=2 00:25:06.183 13:45:13 spdkcli_tcp -- scripts/common.sh@341 -- # ver2_l=1 00:25:06.183 13:45:13 spdkcli_tcp -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:25:06.183 13:45:13 spdkcli_tcp -- scripts/common.sh@344 -- # case "$op" in 00:25:06.183 13:45:13 spdkcli_tcp -- scripts/common.sh@345 -- # : 1 00:25:06.183 13:45:13 spdkcli_tcp -- scripts/common.sh@364 -- # (( v = 0 )) 00:25:06.183 13:45:13 spdkcli_tcp -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:25:06.183 13:45:13 spdkcli_tcp -- scripts/common.sh@365 -- # decimal 1 00:25:06.183 13:45:13 spdkcli_tcp -- scripts/common.sh@353 -- # local d=1 00:25:06.183 13:45:13 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:25:06.183 13:45:13 spdkcli_tcp -- scripts/common.sh@355 -- # echo 1 00:25:06.183 13:45:13 spdkcli_tcp -- scripts/common.sh@365 -- # ver1[v]=1 00:25:06.183 13:45:13 spdkcli_tcp -- scripts/common.sh@366 -- # decimal 2 00:25:06.183 13:45:13 spdkcli_tcp -- scripts/common.sh@353 -- # local d=2 00:25:06.183 13:45:13 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:25:06.183 13:45:13 spdkcli_tcp -- scripts/common.sh@355 -- # echo 2 00:25:06.183 13:45:13 spdkcli_tcp -- scripts/common.sh@366 -- # ver2[v]=2 00:25:06.183 13:45:13 spdkcli_tcp -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:25:06.183 13:45:13 spdkcli_tcp -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:25:06.183 13:45:13 spdkcli_tcp -- scripts/common.sh@368 -- # return 0 00:25:06.183 13:45:13 spdkcli_tcp -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:25:06.183 13:45:13 spdkcli_tcp -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:25:06.183 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:06.183 --rc genhtml_branch_coverage=1 00:25:06.183 --rc genhtml_function_coverage=1 00:25:06.183 --rc genhtml_legend=1 00:25:06.183 --rc geninfo_all_blocks=1 00:25:06.183 --rc geninfo_unexecuted_blocks=1 00:25:06.183 00:25:06.183 ' 00:25:06.183 13:45:13 spdkcli_tcp -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:25:06.183 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:06.183 --rc genhtml_branch_coverage=1 00:25:06.183 --rc genhtml_function_coverage=1 00:25:06.183 --rc genhtml_legend=1 00:25:06.183 --rc geninfo_all_blocks=1 00:25:06.183 --rc geninfo_unexecuted_blocks=1 00:25:06.183 00:25:06.183 ' 00:25:06.183 13:45:13 spdkcli_tcp -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:25:06.183 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:06.183 --rc genhtml_branch_coverage=1 00:25:06.183 --rc genhtml_function_coverage=1 00:25:06.183 --rc genhtml_legend=1 00:25:06.183 --rc geninfo_all_blocks=1 00:25:06.183 --rc geninfo_unexecuted_blocks=1 00:25:06.183 00:25:06.183 ' 00:25:06.183 13:45:13 spdkcli_tcp -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:25:06.183 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:06.183 --rc genhtml_branch_coverage=1 00:25:06.183 --rc genhtml_function_coverage=1 00:25:06.183 --rc genhtml_legend=1 00:25:06.183 --rc geninfo_all_blocks=1 00:25:06.183 --rc geninfo_unexecuted_blocks=1 00:25:06.183 00:25:06.183 ' 00:25:06.183 13:45:13 spdkcli_tcp -- spdkcli/tcp.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:25:06.183 13:45:13 spdkcli_tcp -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:25:06.183 13:45:13 spdkcli_tcp -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:25:06.183 13:45:13 spdkcli_tcp -- spdkcli/tcp.sh@18 -- # IP_ADDRESS=127.0.0.1 00:25:06.183 13:45:13 spdkcli_tcp -- spdkcli/tcp.sh@19 -- # PORT=9998 00:25:06.183 13:45:13 spdkcli_tcp -- spdkcli/tcp.sh@21 -- # trap 'err_cleanup; exit 1' SIGINT SIGTERM EXIT 00:25:06.183 13:45:13 spdkcli_tcp -- spdkcli/tcp.sh@23 -- # timing_enter run_spdk_tgt_tcp 00:25:06.183 13:45:13 spdkcli_tcp -- common/autotest_common.sh@726 -- # xtrace_disable 00:25:06.183 13:45:13 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:25:06.183 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:06.183 13:45:13 spdkcli_tcp -- spdkcli/tcp.sh@25 -- # spdk_tgt_pid=56983 00:25:06.183 13:45:13 spdkcli_tcp -- spdkcli/tcp.sh@27 -- # waitforlisten 56983 00:25:06.183 13:45:13 spdkcli_tcp -- common/autotest_common.sh@835 -- # '[' -z 56983 ']' 00:25:06.183 13:45:13 spdkcli_tcp -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:06.183 13:45:13 spdkcli_tcp -- common/autotest_common.sh@840 -- # local max_retries=100 00:25:06.183 13:45:13 spdkcli_tcp -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:06.183 13:45:13 spdkcli_tcp -- common/autotest_common.sh@844 -- # xtrace_disable 00:25:06.183 13:45:13 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:25:06.183 13:45:13 spdkcli_tcp -- spdkcli/tcp.sh@24 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:25:06.183 [2024-11-20 13:45:13.091128] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:25:06.183 [2024-11-20 13:45:13.091262] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid56983 ] 00:25:06.440 [2024-11-20 13:45:13.251394] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:25:06.440 [2024-11-20 13:45:13.371563] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:25:06.440 [2024-11-20 13:45:13.371662] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:07.005 13:45:14 spdkcli_tcp -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:25:07.005 13:45:14 spdkcli_tcp -- common/autotest_common.sh@868 -- # return 0 00:25:07.005 13:45:14 spdkcli_tcp -- spdkcli/tcp.sh@31 -- # socat_pid=57000 00:25:07.005 13:45:14 spdkcli_tcp -- spdkcli/tcp.sh@33 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -r 100 -t 2 -s 127.0.0.1 -p 9998 rpc_get_methods 00:25:07.005 13:45:14 spdkcli_tcp -- spdkcli/tcp.sh@30 -- # socat TCP-LISTEN:9998 UNIX-CONNECT:/var/tmp/spdk.sock 00:25:07.269 [ 00:25:07.269 "bdev_malloc_delete", 00:25:07.269 "bdev_malloc_create", 00:25:07.269 "bdev_null_resize", 00:25:07.269 "bdev_null_delete", 00:25:07.269 "bdev_null_create", 00:25:07.269 "bdev_nvme_cuse_unregister", 00:25:07.269 "bdev_nvme_cuse_register", 00:25:07.269 "bdev_opal_new_user", 00:25:07.269 "bdev_opal_set_lock_state", 00:25:07.269 "bdev_opal_delete", 00:25:07.269 "bdev_opal_get_info", 00:25:07.269 "bdev_opal_create", 00:25:07.269 "bdev_nvme_opal_revert", 00:25:07.269 "bdev_nvme_opal_init", 00:25:07.269 "bdev_nvme_send_cmd", 00:25:07.269 "bdev_nvme_set_keys", 00:25:07.269 "bdev_nvme_get_path_iostat", 00:25:07.269 "bdev_nvme_get_mdns_discovery_info", 00:25:07.269 "bdev_nvme_stop_mdns_discovery", 00:25:07.269 "bdev_nvme_start_mdns_discovery", 00:25:07.269 "bdev_nvme_set_multipath_policy", 00:25:07.269 "bdev_nvme_set_preferred_path", 00:25:07.269 "bdev_nvme_get_io_paths", 00:25:07.269 "bdev_nvme_remove_error_injection", 00:25:07.269 "bdev_nvme_add_error_injection", 00:25:07.269 "bdev_nvme_get_discovery_info", 00:25:07.269 "bdev_nvme_stop_discovery", 00:25:07.269 "bdev_nvme_start_discovery", 00:25:07.269 "bdev_nvme_get_controller_health_info", 00:25:07.269 "bdev_nvme_disable_controller", 00:25:07.269 "bdev_nvme_enable_controller", 00:25:07.269 "bdev_nvme_reset_controller", 00:25:07.269 "bdev_nvme_get_transport_statistics", 00:25:07.269 "bdev_nvme_apply_firmware", 00:25:07.269 "bdev_nvme_detach_controller", 00:25:07.269 "bdev_nvme_get_controllers", 00:25:07.269 "bdev_nvme_attach_controller", 00:25:07.269 "bdev_nvme_set_hotplug", 00:25:07.269 "bdev_nvme_set_options", 00:25:07.269 "bdev_passthru_delete", 00:25:07.269 "bdev_passthru_create", 00:25:07.269 "bdev_lvol_set_parent_bdev", 00:25:07.269 "bdev_lvol_set_parent", 00:25:07.269 "bdev_lvol_check_shallow_copy", 00:25:07.269 "bdev_lvol_start_shallow_copy", 00:25:07.269 "bdev_lvol_grow_lvstore", 00:25:07.269 "bdev_lvol_get_lvols", 00:25:07.269 "bdev_lvol_get_lvstores", 00:25:07.269 "bdev_lvol_delete", 00:25:07.269 "bdev_lvol_set_read_only", 00:25:07.269 "bdev_lvol_resize", 00:25:07.269 "bdev_lvol_decouple_parent", 00:25:07.269 "bdev_lvol_inflate", 00:25:07.269 "bdev_lvol_rename", 00:25:07.269 "bdev_lvol_clone_bdev", 00:25:07.269 "bdev_lvol_clone", 00:25:07.269 "bdev_lvol_snapshot", 00:25:07.269 "bdev_lvol_create", 00:25:07.269 "bdev_lvol_delete_lvstore", 00:25:07.269 "bdev_lvol_rename_lvstore", 00:25:07.269 "bdev_lvol_create_lvstore", 00:25:07.269 "bdev_raid_set_options", 00:25:07.269 "bdev_raid_remove_base_bdev", 00:25:07.269 "bdev_raid_add_base_bdev", 00:25:07.269 "bdev_raid_delete", 00:25:07.269 "bdev_raid_create", 00:25:07.269 "bdev_raid_get_bdevs", 00:25:07.269 "bdev_error_inject_error", 00:25:07.269 "bdev_error_delete", 00:25:07.269 "bdev_error_create", 00:25:07.269 "bdev_split_delete", 00:25:07.269 "bdev_split_create", 00:25:07.269 "bdev_delay_delete", 00:25:07.269 "bdev_delay_create", 00:25:07.269 "bdev_delay_update_latency", 00:25:07.269 "bdev_zone_block_delete", 00:25:07.269 "bdev_zone_block_create", 00:25:07.269 "blobfs_create", 00:25:07.269 "blobfs_detect", 00:25:07.270 "blobfs_set_cache_size", 00:25:07.270 "bdev_aio_delete", 00:25:07.270 "bdev_aio_rescan", 00:25:07.270 "bdev_aio_create", 00:25:07.270 "bdev_ftl_set_property", 00:25:07.270 "bdev_ftl_get_properties", 00:25:07.270 "bdev_ftl_get_stats", 00:25:07.270 "bdev_ftl_unmap", 00:25:07.270 "bdev_ftl_unload", 00:25:07.270 "bdev_ftl_delete", 00:25:07.270 "bdev_ftl_load", 00:25:07.270 "bdev_ftl_create", 00:25:07.270 "bdev_virtio_attach_controller", 00:25:07.270 "bdev_virtio_scsi_get_devices", 00:25:07.270 "bdev_virtio_detach_controller", 00:25:07.270 "bdev_virtio_blk_set_hotplug", 00:25:07.270 "bdev_iscsi_delete", 00:25:07.270 "bdev_iscsi_create", 00:25:07.270 "bdev_iscsi_set_options", 00:25:07.270 "accel_error_inject_error", 00:25:07.270 "ioat_scan_accel_module", 00:25:07.270 "dsa_scan_accel_module", 00:25:07.270 "iaa_scan_accel_module", 00:25:07.270 "keyring_file_remove_key", 00:25:07.270 "keyring_file_add_key", 00:25:07.270 "keyring_linux_set_options", 00:25:07.270 "fsdev_aio_delete", 00:25:07.270 "fsdev_aio_create", 00:25:07.270 "iscsi_get_histogram", 00:25:07.270 "iscsi_enable_histogram", 00:25:07.270 "iscsi_set_options", 00:25:07.270 "iscsi_get_auth_groups", 00:25:07.270 "iscsi_auth_group_remove_secret", 00:25:07.270 "iscsi_auth_group_add_secret", 00:25:07.270 "iscsi_delete_auth_group", 00:25:07.270 "iscsi_create_auth_group", 00:25:07.270 "iscsi_set_discovery_auth", 00:25:07.270 "iscsi_get_options", 00:25:07.270 "iscsi_target_node_request_logout", 00:25:07.270 "iscsi_target_node_set_redirect", 00:25:07.270 "iscsi_target_node_set_auth", 00:25:07.270 "iscsi_target_node_add_lun", 00:25:07.270 "iscsi_get_stats", 00:25:07.270 "iscsi_get_connections", 00:25:07.270 "iscsi_portal_group_set_auth", 00:25:07.270 "iscsi_start_portal_group", 00:25:07.270 "iscsi_delete_portal_group", 00:25:07.270 "iscsi_create_portal_group", 00:25:07.270 "iscsi_get_portal_groups", 00:25:07.270 "iscsi_delete_target_node", 00:25:07.270 "iscsi_target_node_remove_pg_ig_maps", 00:25:07.270 "iscsi_target_node_add_pg_ig_maps", 00:25:07.270 "iscsi_create_target_node", 00:25:07.270 "iscsi_get_target_nodes", 00:25:07.270 "iscsi_delete_initiator_group", 00:25:07.270 "iscsi_initiator_group_remove_initiators", 00:25:07.270 "iscsi_initiator_group_add_initiators", 00:25:07.270 "iscsi_create_initiator_group", 00:25:07.270 "iscsi_get_initiator_groups", 00:25:07.270 "nvmf_set_crdt", 00:25:07.270 "nvmf_set_config", 00:25:07.270 "nvmf_set_max_subsystems", 00:25:07.270 "nvmf_stop_mdns_prr", 00:25:07.270 "nvmf_publish_mdns_prr", 00:25:07.270 "nvmf_subsystem_get_listeners", 00:25:07.270 "nvmf_subsystem_get_qpairs", 00:25:07.270 "nvmf_subsystem_get_controllers", 00:25:07.270 "nvmf_get_stats", 00:25:07.270 "nvmf_get_transports", 00:25:07.270 "nvmf_create_transport", 00:25:07.270 "nvmf_get_targets", 00:25:07.270 "nvmf_delete_target", 00:25:07.270 "nvmf_create_target", 00:25:07.270 "nvmf_subsystem_allow_any_host", 00:25:07.270 "nvmf_subsystem_set_keys", 00:25:07.270 "nvmf_subsystem_remove_host", 00:25:07.270 "nvmf_subsystem_add_host", 00:25:07.270 "nvmf_ns_remove_host", 00:25:07.270 "nvmf_ns_add_host", 00:25:07.270 "nvmf_subsystem_remove_ns", 00:25:07.270 "nvmf_subsystem_set_ns_ana_group", 00:25:07.270 "nvmf_subsystem_add_ns", 00:25:07.270 "nvmf_subsystem_listener_set_ana_state", 00:25:07.270 "nvmf_discovery_get_referrals", 00:25:07.270 "nvmf_discovery_remove_referral", 00:25:07.270 "nvmf_discovery_add_referral", 00:25:07.270 "nvmf_subsystem_remove_listener", 00:25:07.270 "nvmf_subsystem_add_listener", 00:25:07.270 "nvmf_delete_subsystem", 00:25:07.270 "nvmf_create_subsystem", 00:25:07.270 "nvmf_get_subsystems", 00:25:07.270 "env_dpdk_get_mem_stats", 00:25:07.270 "nbd_get_disks", 00:25:07.270 "nbd_stop_disk", 00:25:07.270 "nbd_start_disk", 00:25:07.270 "ublk_recover_disk", 00:25:07.270 "ublk_get_disks", 00:25:07.270 "ublk_stop_disk", 00:25:07.270 "ublk_start_disk", 00:25:07.270 "ublk_destroy_target", 00:25:07.270 "ublk_create_target", 00:25:07.270 "virtio_blk_create_transport", 00:25:07.270 "virtio_blk_get_transports", 00:25:07.270 "vhost_controller_set_coalescing", 00:25:07.270 "vhost_get_controllers", 00:25:07.270 "vhost_delete_controller", 00:25:07.270 "vhost_create_blk_controller", 00:25:07.270 "vhost_scsi_controller_remove_target", 00:25:07.270 "vhost_scsi_controller_add_target", 00:25:07.270 "vhost_start_scsi_controller", 00:25:07.270 "vhost_create_scsi_controller", 00:25:07.270 "thread_set_cpumask", 00:25:07.270 "scheduler_set_options", 00:25:07.270 "framework_get_governor", 00:25:07.270 "framework_get_scheduler", 00:25:07.270 "framework_set_scheduler", 00:25:07.270 "framework_get_reactors", 00:25:07.270 "thread_get_io_channels", 00:25:07.270 "thread_get_pollers", 00:25:07.270 "thread_get_stats", 00:25:07.270 "framework_monitor_context_switch", 00:25:07.270 "spdk_kill_instance", 00:25:07.270 "log_enable_timestamps", 00:25:07.270 "log_get_flags", 00:25:07.270 "log_clear_flag", 00:25:07.270 "log_set_flag", 00:25:07.270 "log_get_level", 00:25:07.270 "log_set_level", 00:25:07.270 "log_get_print_level", 00:25:07.270 "log_set_print_level", 00:25:07.270 "framework_enable_cpumask_locks", 00:25:07.270 "framework_disable_cpumask_locks", 00:25:07.270 "framework_wait_init", 00:25:07.270 "framework_start_init", 00:25:07.270 "scsi_get_devices", 00:25:07.270 "bdev_get_histogram", 00:25:07.270 "bdev_enable_histogram", 00:25:07.270 "bdev_set_qos_limit", 00:25:07.270 "bdev_set_qd_sampling_period", 00:25:07.270 "bdev_get_bdevs", 00:25:07.270 "bdev_reset_iostat", 00:25:07.270 "bdev_get_iostat", 00:25:07.270 "bdev_examine", 00:25:07.270 "bdev_wait_for_examine", 00:25:07.270 "bdev_set_options", 00:25:07.270 "accel_get_stats", 00:25:07.270 "accel_set_options", 00:25:07.270 "accel_set_driver", 00:25:07.270 "accel_crypto_key_destroy", 00:25:07.270 "accel_crypto_keys_get", 00:25:07.270 "accel_crypto_key_create", 00:25:07.270 "accel_assign_opc", 00:25:07.270 "accel_get_module_info", 00:25:07.270 "accel_get_opc_assignments", 00:25:07.270 "vmd_rescan", 00:25:07.270 "vmd_remove_device", 00:25:07.270 "vmd_enable", 00:25:07.270 "sock_get_default_impl", 00:25:07.270 "sock_set_default_impl", 00:25:07.270 "sock_impl_set_options", 00:25:07.270 "sock_impl_get_options", 00:25:07.270 "iobuf_get_stats", 00:25:07.270 "iobuf_set_options", 00:25:07.270 "keyring_get_keys", 00:25:07.270 "framework_get_pci_devices", 00:25:07.270 "framework_get_config", 00:25:07.270 "framework_get_subsystems", 00:25:07.270 "fsdev_set_opts", 00:25:07.270 "fsdev_get_opts", 00:25:07.270 "trace_get_info", 00:25:07.270 "trace_get_tpoint_group_mask", 00:25:07.270 "trace_disable_tpoint_group", 00:25:07.270 "trace_enable_tpoint_group", 00:25:07.271 "trace_clear_tpoint_mask", 00:25:07.271 "trace_set_tpoint_mask", 00:25:07.271 "notify_get_notifications", 00:25:07.271 "notify_get_types", 00:25:07.271 "spdk_get_version", 00:25:07.271 "rpc_get_methods" 00:25:07.271 ] 00:25:07.271 13:45:14 spdkcli_tcp -- spdkcli/tcp.sh@35 -- # timing_exit run_spdk_tgt_tcp 00:25:07.271 13:45:14 spdkcli_tcp -- common/autotest_common.sh@732 -- # xtrace_disable 00:25:07.271 13:45:14 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:25:07.271 13:45:14 spdkcli_tcp -- spdkcli/tcp.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:25:07.271 13:45:14 spdkcli_tcp -- spdkcli/tcp.sh@38 -- # killprocess 56983 00:25:07.271 13:45:14 spdkcli_tcp -- common/autotest_common.sh@954 -- # '[' -z 56983 ']' 00:25:07.271 13:45:14 spdkcli_tcp -- common/autotest_common.sh@958 -- # kill -0 56983 00:25:07.271 13:45:14 spdkcli_tcp -- common/autotest_common.sh@959 -- # uname 00:25:07.271 13:45:14 spdkcli_tcp -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:25:07.271 13:45:14 spdkcli_tcp -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 56983 00:25:07.271 killing process with pid 56983 00:25:07.271 13:45:14 spdkcli_tcp -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:25:07.271 13:45:14 spdkcli_tcp -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:25:07.271 13:45:14 spdkcli_tcp -- common/autotest_common.sh@972 -- # echo 'killing process with pid 56983' 00:25:07.271 13:45:14 spdkcli_tcp -- common/autotest_common.sh@973 -- # kill 56983 00:25:07.271 13:45:14 spdkcli_tcp -- common/autotest_common.sh@978 -- # wait 56983 00:25:09.169 ************************************ 00:25:09.169 END TEST spdkcli_tcp 00:25:09.169 ************************************ 00:25:09.169 00:25:09.169 real 0m2.943s 00:25:09.169 user 0m5.329s 00:25:09.169 sys 0m0.456s 00:25:09.169 13:45:15 spdkcli_tcp -- common/autotest_common.sh@1130 -- # xtrace_disable 00:25:09.169 13:45:15 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:25:09.169 13:45:15 -- spdk/autotest.sh@167 -- # run_test dpdk_mem_utility /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:25:09.169 13:45:15 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:25:09.169 13:45:15 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:25:09.169 13:45:15 -- common/autotest_common.sh@10 -- # set +x 00:25:09.169 ************************************ 00:25:09.169 START TEST dpdk_mem_utility 00:25:09.169 ************************************ 00:25:09.170 13:45:15 dpdk_mem_utility -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:25:09.170 * Looking for test storage... 00:25:09.170 * Found test storage at /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility 00:25:09.170 13:45:15 dpdk_mem_utility -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:25:09.170 13:45:15 dpdk_mem_utility -- common/autotest_common.sh@1693 -- # lcov --version 00:25:09.170 13:45:15 dpdk_mem_utility -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:25:09.170 13:45:15 dpdk_mem_utility -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:25:09.170 13:45:15 dpdk_mem_utility -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:25:09.170 13:45:15 dpdk_mem_utility -- scripts/common.sh@333 -- # local ver1 ver1_l 00:25:09.170 13:45:15 dpdk_mem_utility -- scripts/common.sh@334 -- # local ver2 ver2_l 00:25:09.170 13:45:15 dpdk_mem_utility -- scripts/common.sh@336 -- # IFS=.-: 00:25:09.170 13:45:15 dpdk_mem_utility -- scripts/common.sh@336 -- # read -ra ver1 00:25:09.170 13:45:15 dpdk_mem_utility -- scripts/common.sh@337 -- # IFS=.-: 00:25:09.170 13:45:15 dpdk_mem_utility -- scripts/common.sh@337 -- # read -ra ver2 00:25:09.170 13:45:15 dpdk_mem_utility -- scripts/common.sh@338 -- # local 'op=<' 00:25:09.170 13:45:15 dpdk_mem_utility -- scripts/common.sh@340 -- # ver1_l=2 00:25:09.170 13:45:15 dpdk_mem_utility -- scripts/common.sh@341 -- # ver2_l=1 00:25:09.170 13:45:15 dpdk_mem_utility -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:25:09.170 13:45:15 dpdk_mem_utility -- scripts/common.sh@344 -- # case "$op" in 00:25:09.170 13:45:15 dpdk_mem_utility -- scripts/common.sh@345 -- # : 1 00:25:09.170 13:45:15 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v = 0 )) 00:25:09.170 13:45:15 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:25:09.170 13:45:15 dpdk_mem_utility -- scripts/common.sh@365 -- # decimal 1 00:25:09.170 13:45:15 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=1 00:25:09.170 13:45:15 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:25:09.170 13:45:15 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 1 00:25:09.170 13:45:15 dpdk_mem_utility -- scripts/common.sh@365 -- # ver1[v]=1 00:25:09.170 13:45:15 dpdk_mem_utility -- scripts/common.sh@366 -- # decimal 2 00:25:09.170 13:45:15 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=2 00:25:09.170 13:45:15 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:25:09.170 13:45:15 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 2 00:25:09.170 13:45:15 dpdk_mem_utility -- scripts/common.sh@366 -- # ver2[v]=2 00:25:09.170 13:45:16 dpdk_mem_utility -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:25:09.170 13:45:16 dpdk_mem_utility -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:25:09.170 13:45:16 dpdk_mem_utility -- scripts/common.sh@368 -- # return 0 00:25:09.170 13:45:16 dpdk_mem_utility -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:25:09.170 13:45:16 dpdk_mem_utility -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:25:09.170 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:09.170 --rc genhtml_branch_coverage=1 00:25:09.170 --rc genhtml_function_coverage=1 00:25:09.170 --rc genhtml_legend=1 00:25:09.170 --rc geninfo_all_blocks=1 00:25:09.170 --rc geninfo_unexecuted_blocks=1 00:25:09.170 00:25:09.170 ' 00:25:09.170 13:45:16 dpdk_mem_utility -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:25:09.170 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:09.170 --rc genhtml_branch_coverage=1 00:25:09.170 --rc genhtml_function_coverage=1 00:25:09.170 --rc genhtml_legend=1 00:25:09.170 --rc geninfo_all_blocks=1 00:25:09.170 --rc geninfo_unexecuted_blocks=1 00:25:09.170 00:25:09.170 ' 00:25:09.170 13:45:16 dpdk_mem_utility -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:25:09.170 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:09.170 --rc genhtml_branch_coverage=1 00:25:09.170 --rc genhtml_function_coverage=1 00:25:09.170 --rc genhtml_legend=1 00:25:09.170 --rc geninfo_all_blocks=1 00:25:09.170 --rc geninfo_unexecuted_blocks=1 00:25:09.170 00:25:09.170 ' 00:25:09.170 13:45:16 dpdk_mem_utility -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:25:09.170 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:09.170 --rc genhtml_branch_coverage=1 00:25:09.170 --rc genhtml_function_coverage=1 00:25:09.170 --rc genhtml_legend=1 00:25:09.170 --rc geninfo_all_blocks=1 00:25:09.170 --rc geninfo_unexecuted_blocks=1 00:25:09.170 00:25:09.170 ' 00:25:09.170 13:45:16 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@10 -- # MEM_SCRIPT=/home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py 00:25:09.170 13:45:16 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@13 -- # spdkpid=57093 00:25:09.170 13:45:16 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@15 -- # waitforlisten 57093 00:25:09.170 13:45:16 dpdk_mem_utility -- common/autotest_common.sh@835 -- # '[' -z 57093 ']' 00:25:09.170 13:45:16 dpdk_mem_utility -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:09.170 13:45:16 dpdk_mem_utility -- common/autotest_common.sh@840 -- # local max_retries=100 00:25:09.170 13:45:16 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@12 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:25:09.170 13:45:16 dpdk_mem_utility -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:09.170 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:09.170 13:45:16 dpdk_mem_utility -- common/autotest_common.sh@844 -- # xtrace_disable 00:25:09.170 13:45:16 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:25:09.170 [2024-11-20 13:45:16.085733] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:25:09.170 [2024-11-20 13:45:16.085940] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57093 ] 00:25:09.427 [2024-11-20 13:45:16.247665] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:09.427 [2024-11-20 13:45:16.365240] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:09.992 13:45:17 dpdk_mem_utility -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:25:09.992 13:45:17 dpdk_mem_utility -- common/autotest_common.sh@868 -- # return 0 00:25:09.992 13:45:17 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@17 -- # trap 'killprocess $spdkpid' SIGINT SIGTERM EXIT 00:25:09.992 13:45:17 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@19 -- # rpc_cmd env_dpdk_get_mem_stats 00:25:09.992 13:45:17 dpdk_mem_utility -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:09.992 13:45:17 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:25:09.992 { 00:25:09.992 "filename": "/tmp/spdk_mem_dump.txt" 00:25:09.992 } 00:25:09.992 13:45:17 dpdk_mem_utility -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:09.992 13:45:17 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@21 -- # /home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py 00:25:10.252 DPDK memory size 824.000000 MiB in 1 heap(s) 00:25:10.252 1 heaps totaling size 824.000000 MiB 00:25:10.252 size: 824.000000 MiB heap id: 0 00:25:10.252 end heaps---------- 00:25:10.252 9 mempools totaling size 603.782043 MiB 00:25:10.252 size: 212.674988 MiB name: PDU_immediate_data_Pool 00:25:10.252 size: 158.602051 MiB name: PDU_data_out_Pool 00:25:10.252 size: 100.555481 MiB name: bdev_io_57093 00:25:10.252 size: 50.003479 MiB name: msgpool_57093 00:25:10.252 size: 36.509338 MiB name: fsdev_io_57093 00:25:10.252 size: 21.763794 MiB name: PDU_Pool 00:25:10.252 size: 19.513306 MiB name: SCSI_TASK_Pool 00:25:10.252 size: 4.133484 MiB name: evtpool_57093 00:25:10.252 size: 0.026123 MiB name: Session_Pool 00:25:10.252 end mempools------- 00:25:10.252 6 memzones totaling size 4.142822 MiB 00:25:10.252 size: 1.000366 MiB name: RG_ring_0_57093 00:25:10.252 size: 1.000366 MiB name: RG_ring_1_57093 00:25:10.252 size: 1.000366 MiB name: RG_ring_4_57093 00:25:10.252 size: 1.000366 MiB name: RG_ring_5_57093 00:25:10.252 size: 0.125366 MiB name: RG_ring_2_57093 00:25:10.252 size: 0.015991 MiB name: RG_ring_3_57093 00:25:10.252 end memzones------- 00:25:10.252 13:45:17 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@23 -- # /home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py -m 0 00:25:10.252 heap id: 0 total size: 824.000000 MiB number of busy elements: 324 number of free elements: 18 00:25:10.252 list of free elements. size: 16.779175 MiB 00:25:10.252 element at address: 0x200006400000 with size: 1.995972 MiB 00:25:10.252 element at address: 0x20000a600000 with size: 1.995972 MiB 00:25:10.252 element at address: 0x200003e00000 with size: 1.991028 MiB 00:25:10.252 element at address: 0x200019500040 with size: 0.999939 MiB 00:25:10.252 element at address: 0x200019900040 with size: 0.999939 MiB 00:25:10.252 element at address: 0x200019a00000 with size: 0.999084 MiB 00:25:10.252 element at address: 0x200032600000 with size: 0.994324 MiB 00:25:10.252 element at address: 0x200000400000 with size: 0.992004 MiB 00:25:10.252 element at address: 0x200019200000 with size: 0.959656 MiB 00:25:10.252 element at address: 0x200019d00040 with size: 0.936401 MiB 00:25:10.252 element at address: 0x200000200000 with size: 0.716980 MiB 00:25:10.252 element at address: 0x20001b400000 with size: 0.559509 MiB 00:25:10.252 element at address: 0x200000c00000 with size: 0.489197 MiB 00:25:10.252 element at address: 0x200019600000 with size: 0.487976 MiB 00:25:10.252 element at address: 0x200019e00000 with size: 0.485413 MiB 00:25:10.252 element at address: 0x200012c00000 with size: 0.433472 MiB 00:25:10.252 element at address: 0x200028800000 with size: 0.391418 MiB 00:25:10.252 element at address: 0x200000800000 with size: 0.350891 MiB 00:25:10.252 list of standard malloc elements. size: 199.289917 MiB 00:25:10.252 element at address: 0x20000a7fef80 with size: 132.000183 MiB 00:25:10.252 element at address: 0x2000065fef80 with size: 64.000183 MiB 00:25:10.252 element at address: 0x2000193fff80 with size: 1.000183 MiB 00:25:10.252 element at address: 0x2000197fff80 with size: 1.000183 MiB 00:25:10.252 element at address: 0x200019bfff80 with size: 1.000183 MiB 00:25:10.252 element at address: 0x2000003d9e80 with size: 0.140808 MiB 00:25:10.252 element at address: 0x200019deff40 with size: 0.062683 MiB 00:25:10.252 element at address: 0x2000003fdf40 with size: 0.007996 MiB 00:25:10.252 element at address: 0x20000a5ff040 with size: 0.000427 MiB 00:25:10.252 element at address: 0x200019defdc0 with size: 0.000366 MiB 00:25:10.252 element at address: 0x200012bff040 with size: 0.000305 MiB 00:25:10.252 element at address: 0x2000002d7b00 with size: 0.000244 MiB 00:25:10.252 element at address: 0x2000003d9d80 with size: 0.000244 MiB 00:25:10.252 element at address: 0x2000004fdf40 with size: 0.000244 MiB 00:25:10.252 element at address: 0x2000004fe040 with size: 0.000244 MiB 00:25:10.252 element at address: 0x2000004fe140 with size: 0.000244 MiB 00:25:10.252 element at address: 0x2000004fe240 with size: 0.000244 MiB 00:25:10.252 element at address: 0x2000004fe340 with size: 0.000244 MiB 00:25:10.252 element at address: 0x2000004fe440 with size: 0.000244 MiB 00:25:10.252 element at address: 0x2000004fe540 with size: 0.000244 MiB 00:25:10.252 element at address: 0x2000004fe640 with size: 0.000244 MiB 00:25:10.252 element at address: 0x2000004fe740 with size: 0.000244 MiB 00:25:10.252 element at address: 0x2000004fe840 with size: 0.000244 MiB 00:25:10.252 element at address: 0x2000004fe940 with size: 0.000244 MiB 00:25:10.252 element at address: 0x2000004fea40 with size: 0.000244 MiB 00:25:10.253 element at address: 0x2000004feb40 with size: 0.000244 MiB 00:25:10.253 element at address: 0x2000004fec40 with size: 0.000244 MiB 00:25:10.253 element at address: 0x2000004fed40 with size: 0.000244 MiB 00:25:10.253 element at address: 0x2000004fee40 with size: 0.000244 MiB 00:25:10.253 element at address: 0x2000004fef40 with size: 0.000244 MiB 00:25:10.253 element at address: 0x2000004ff040 with size: 0.000244 MiB 00:25:10.253 element at address: 0x2000004ff140 with size: 0.000244 MiB 00:25:10.253 element at address: 0x2000004ff240 with size: 0.000244 MiB 00:25:10.253 element at address: 0x2000004ff340 with size: 0.000244 MiB 00:25:10.253 element at address: 0x2000004ff440 with size: 0.000244 MiB 00:25:10.253 element at address: 0x2000004ff540 with size: 0.000244 MiB 00:25:10.253 element at address: 0x2000004ff640 with size: 0.000244 MiB 00:25:10.253 element at address: 0x2000004ff740 with size: 0.000244 MiB 00:25:10.253 element at address: 0x2000004ff840 with size: 0.000244 MiB 00:25:10.253 element at address: 0x2000004ff940 with size: 0.000244 MiB 00:25:10.253 element at address: 0x2000004ffbc0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x2000004ffcc0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x2000004ffdc0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000087e1c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000087e2c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000087e3c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000087e4c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000087e5c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000087e6c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000087e7c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000087e8c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000087e9c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000087eac0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000087ebc0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000087ecc0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000087edc0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000087eec0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000087efc0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000087f0c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000087f1c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000087f2c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000087f3c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000087f4c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x2000008ff800 with size: 0.000244 MiB 00:25:10.253 element at address: 0x2000008ffa80 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000c7d3c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000c7d4c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000c7d5c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000c7d6c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000c7d7c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000c7d8c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000c7d9c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000c7dac0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000c7dbc0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000c7dcc0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000c7ddc0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000c7dec0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000c7dfc0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000c7e0c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000c7e1c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000c7e2c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000c7e3c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000c7e4c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000c7e5c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000c7e6c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000c7e7c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000c7e8c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000c7e9c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000c7eac0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000c7ebc0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000cfef00 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200000cff000 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000a5ff200 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000a5ff300 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000a5ff400 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000a5ff500 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000a5ff600 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000a5ff700 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000a5ff800 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000a5ff900 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000a5ffa00 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000a5ffb00 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000a5ffc00 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000a5ffd00 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000a5ffe00 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20000a5fff00 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200012bff180 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200012bff280 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200012bff380 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200012bff480 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200012bff580 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200012bff680 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200012bff780 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200012bff880 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200012bff980 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200012bffa80 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200012bffb80 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200012bffc80 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200012bfff00 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200012c6ef80 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200012c6f080 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200012c6f180 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200012c6f280 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200012c6f380 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200012c6f480 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200012c6f580 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200012c6f680 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200012c6f780 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200012c6f880 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200012cefbc0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x2000192fdd00 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001967cec0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001967cfc0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001967d0c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001967d1c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001967d2c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001967d3c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001967d4c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001967d5c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001967d6c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001967d7c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001967d8c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001967d9c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x2000196fdd00 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200019affc40 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200019defbc0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200019defcc0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x200019ebc680 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b48f3c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b48f4c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b48f5c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b48f6c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b48f7c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b48f8c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b48f9c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b48fac0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b48fbc0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b48fcc0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b48fdc0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b48fec0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b48ffc0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b4900c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b4901c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b4902c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b4903c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b4904c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b4905c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b4906c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b4907c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b4908c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b4909c0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b490ac0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b490bc0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b490cc0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b490dc0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b490ec0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b490fc0 with size: 0.000244 MiB 00:25:10.253 element at address: 0x20001b4910c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4911c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4912c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4913c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4914c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4915c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4916c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4917c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4918c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4919c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b491ac0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b491bc0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b491cc0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b491dc0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b491ec0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b491fc0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4920c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4921c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4922c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4923c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4924c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4925c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4926c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4927c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4928c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4929c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b492ac0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b492bc0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b492cc0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b492dc0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b492ec0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b492fc0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4930c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4931c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4932c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4933c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4934c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4935c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4936c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4937c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4938c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4939c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b493ac0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b493bc0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b493cc0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b493dc0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b493ec0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b493fc0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4940c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4941c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4942c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4943c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4944c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4945c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4946c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4947c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4948c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4949c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b494ac0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b494bc0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b494cc0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b494dc0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b494ec0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b494fc0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4950c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4951c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4952c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20001b4953c0 with size: 0.000244 MiB 00:25:10.254 element at address: 0x200028864340 with size: 0.000244 MiB 00:25:10.254 element at address: 0x200028864440 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886b100 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886b380 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886b480 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886b580 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886b680 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886b780 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886b880 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886b980 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886ba80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886bb80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886bc80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886bd80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886be80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886bf80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886c080 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886c180 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886c280 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886c380 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886c480 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886c580 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886c680 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886c780 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886c880 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886c980 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886ca80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886cb80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886cc80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886cd80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886ce80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886cf80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886d080 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886d180 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886d280 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886d380 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886d480 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886d580 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886d680 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886d780 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886d880 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886d980 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886da80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886db80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886dc80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886dd80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886de80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886df80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886e080 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886e180 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886e280 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886e380 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886e480 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886e580 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886e680 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886e780 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886e880 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886e980 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886ea80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886eb80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886ec80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886ed80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886ee80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886ef80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886f080 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886f180 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886f280 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886f380 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886f480 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886f580 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886f680 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886f780 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886f880 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886f980 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886fa80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886fb80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886fc80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886fd80 with size: 0.000244 MiB 00:25:10.254 element at address: 0x20002886fe80 with size: 0.000244 MiB 00:25:10.254 list of memzone associated elements. size: 607.930908 MiB 00:25:10.254 element at address: 0x20001b4954c0 with size: 211.416809 MiB 00:25:10.254 associated memzone info: size: 211.416626 MiB name: MP_PDU_immediate_data_Pool_0 00:25:10.254 element at address: 0x20002886ff80 with size: 157.562622 MiB 00:25:10.255 associated memzone info: size: 157.562439 MiB name: MP_PDU_data_out_Pool_0 00:25:10.255 element at address: 0x200012df1e40 with size: 100.055115 MiB 00:25:10.255 associated memzone info: size: 100.054932 MiB name: MP_bdev_io_57093_0 00:25:10.255 element at address: 0x200000dff340 with size: 48.003113 MiB 00:25:10.255 associated memzone info: size: 48.002930 MiB name: MP_msgpool_57093_0 00:25:10.255 element at address: 0x200003ffdb40 with size: 36.008972 MiB 00:25:10.255 associated memzone info: size: 36.008789 MiB name: MP_fsdev_io_57093_0 00:25:10.255 element at address: 0x200019fbe900 with size: 20.255615 MiB 00:25:10.255 associated memzone info: size: 20.255432 MiB name: MP_PDU_Pool_0 00:25:10.255 element at address: 0x2000327feb00 with size: 18.005127 MiB 00:25:10.255 associated memzone info: size: 18.004944 MiB name: MP_SCSI_TASK_Pool_0 00:25:10.255 element at address: 0x2000004ffec0 with size: 3.000305 MiB 00:25:10.255 associated memzone info: size: 3.000122 MiB name: MP_evtpool_57093_0 00:25:10.255 element at address: 0x2000009ffdc0 with size: 2.000549 MiB 00:25:10.255 associated memzone info: size: 2.000366 MiB name: RG_MP_msgpool_57093 00:25:10.255 element at address: 0x2000002d7c00 with size: 1.008179 MiB 00:25:10.255 associated memzone info: size: 1.007996 MiB name: MP_evtpool_57093 00:25:10.255 element at address: 0x2000196fde00 with size: 1.008179 MiB 00:25:10.255 associated memzone info: size: 1.007996 MiB name: MP_PDU_Pool 00:25:10.255 element at address: 0x200019ebc780 with size: 1.008179 MiB 00:25:10.255 associated memzone info: size: 1.007996 MiB name: MP_PDU_immediate_data_Pool 00:25:10.255 element at address: 0x2000192fde00 with size: 1.008179 MiB 00:25:10.255 associated memzone info: size: 1.007996 MiB name: MP_PDU_data_out_Pool 00:25:10.255 element at address: 0x200012cefcc0 with size: 1.008179 MiB 00:25:10.255 associated memzone info: size: 1.007996 MiB name: MP_SCSI_TASK_Pool 00:25:10.255 element at address: 0x200000cff100 with size: 1.000549 MiB 00:25:10.255 associated memzone info: size: 1.000366 MiB name: RG_ring_0_57093 00:25:10.255 element at address: 0x2000008ffb80 with size: 1.000549 MiB 00:25:10.255 associated memzone info: size: 1.000366 MiB name: RG_ring_1_57093 00:25:10.255 element at address: 0x200019affd40 with size: 1.000549 MiB 00:25:10.255 associated memzone info: size: 1.000366 MiB name: RG_ring_4_57093 00:25:10.255 element at address: 0x2000326fe8c0 with size: 1.000549 MiB 00:25:10.255 associated memzone info: size: 1.000366 MiB name: RG_ring_5_57093 00:25:10.255 element at address: 0x20000087f5c0 with size: 0.500549 MiB 00:25:10.255 associated memzone info: size: 0.500366 MiB name: RG_MP_fsdev_io_57093 00:25:10.255 element at address: 0x200000c7ecc0 with size: 0.500549 MiB 00:25:10.255 associated memzone info: size: 0.500366 MiB name: RG_MP_bdev_io_57093 00:25:10.255 element at address: 0x20001967dac0 with size: 0.500549 MiB 00:25:10.255 associated memzone info: size: 0.500366 MiB name: RG_MP_PDU_Pool 00:25:10.255 element at address: 0x200012c6f980 with size: 0.500549 MiB 00:25:10.255 associated memzone info: size: 0.500366 MiB name: RG_MP_SCSI_TASK_Pool 00:25:10.255 element at address: 0x200019e7c440 with size: 0.250549 MiB 00:25:10.255 associated memzone info: size: 0.250366 MiB name: RG_MP_PDU_immediate_data_Pool 00:25:10.255 element at address: 0x2000002b78c0 with size: 0.125549 MiB 00:25:10.255 associated memzone info: size: 0.125366 MiB name: RG_MP_evtpool_57093 00:25:10.255 element at address: 0x20000085df80 with size: 0.125549 MiB 00:25:10.255 associated memzone info: size: 0.125366 MiB name: RG_ring_2_57093 00:25:10.255 element at address: 0x2000192f5ac0 with size: 0.031799 MiB 00:25:10.255 associated memzone info: size: 0.031616 MiB name: RG_MP_PDU_data_out_Pool 00:25:10.255 element at address: 0x200028864540 with size: 0.023804 MiB 00:25:10.255 associated memzone info: size: 0.023621 MiB name: MP_Session_Pool_0 00:25:10.255 element at address: 0x200000859d40 with size: 0.016174 MiB 00:25:10.255 associated memzone info: size: 0.015991 MiB name: RG_ring_3_57093 00:25:10.255 element at address: 0x20002886a6c0 with size: 0.002502 MiB 00:25:10.255 associated memzone info: size: 0.002319 MiB name: RG_MP_Session_Pool 00:25:10.255 element at address: 0x2000004ffa40 with size: 0.000366 MiB 00:25:10.255 associated memzone info: size: 0.000183 MiB name: MP_msgpool_57093 00:25:10.255 element at address: 0x2000008ff900 with size: 0.000366 MiB 00:25:10.255 associated memzone info: size: 0.000183 MiB name: MP_fsdev_io_57093 00:25:10.255 element at address: 0x200012bffd80 with size: 0.000366 MiB 00:25:10.255 associated memzone info: size: 0.000183 MiB name: MP_bdev_io_57093 00:25:10.255 element at address: 0x20002886b200 with size: 0.000366 MiB 00:25:10.255 associated memzone info: size: 0.000183 MiB name: MP_Session_Pool 00:25:10.255 13:45:17 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@25 -- # trap - SIGINT SIGTERM EXIT 00:25:10.255 13:45:17 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@26 -- # killprocess 57093 00:25:10.255 13:45:17 dpdk_mem_utility -- common/autotest_common.sh@954 -- # '[' -z 57093 ']' 00:25:10.255 13:45:17 dpdk_mem_utility -- common/autotest_common.sh@958 -- # kill -0 57093 00:25:10.255 13:45:17 dpdk_mem_utility -- common/autotest_common.sh@959 -- # uname 00:25:10.255 13:45:17 dpdk_mem_utility -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:25:10.255 13:45:17 dpdk_mem_utility -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 57093 00:25:10.255 13:45:17 dpdk_mem_utility -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:25:10.255 13:45:17 dpdk_mem_utility -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:25:10.255 killing process with pid 57093 00:25:10.255 13:45:17 dpdk_mem_utility -- common/autotest_common.sh@972 -- # echo 'killing process with pid 57093' 00:25:10.255 13:45:17 dpdk_mem_utility -- common/autotest_common.sh@973 -- # kill 57093 00:25:10.255 13:45:17 dpdk_mem_utility -- common/autotest_common.sh@978 -- # wait 57093 00:25:11.628 ************************************ 00:25:11.628 END TEST dpdk_mem_utility 00:25:11.628 00:25:11.628 real 0m2.785s 00:25:11.628 user 0m2.725s 00:25:11.628 sys 0m0.472s 00:25:11.628 13:45:18 dpdk_mem_utility -- common/autotest_common.sh@1130 -- # xtrace_disable 00:25:11.628 13:45:18 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:25:11.628 ************************************ 00:25:11.628 13:45:18 -- spdk/autotest.sh@168 -- # run_test event /home/vagrant/spdk_repo/spdk/test/event/event.sh 00:25:11.628 13:45:18 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:25:11.628 13:45:18 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:25:11.628 13:45:18 -- common/autotest_common.sh@10 -- # set +x 00:25:11.886 ************************************ 00:25:11.886 START TEST event 00:25:11.886 ************************************ 00:25:11.886 13:45:18 event -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/event.sh 00:25:11.886 * Looking for test storage... 00:25:11.886 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event 00:25:11.886 13:45:18 event -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:25:11.886 13:45:18 event -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:25:11.886 13:45:18 event -- common/autotest_common.sh@1693 -- # lcov --version 00:25:11.886 13:45:18 event -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:25:11.886 13:45:18 event -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:25:11.886 13:45:18 event -- scripts/common.sh@333 -- # local ver1 ver1_l 00:25:11.886 13:45:18 event -- scripts/common.sh@334 -- # local ver2 ver2_l 00:25:11.886 13:45:18 event -- scripts/common.sh@336 -- # IFS=.-: 00:25:11.886 13:45:18 event -- scripts/common.sh@336 -- # read -ra ver1 00:25:11.886 13:45:18 event -- scripts/common.sh@337 -- # IFS=.-: 00:25:11.886 13:45:18 event -- scripts/common.sh@337 -- # read -ra ver2 00:25:11.886 13:45:18 event -- scripts/common.sh@338 -- # local 'op=<' 00:25:11.886 13:45:18 event -- scripts/common.sh@340 -- # ver1_l=2 00:25:11.886 13:45:18 event -- scripts/common.sh@341 -- # ver2_l=1 00:25:11.886 13:45:18 event -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:25:11.886 13:45:18 event -- scripts/common.sh@344 -- # case "$op" in 00:25:11.886 13:45:18 event -- scripts/common.sh@345 -- # : 1 00:25:11.886 13:45:18 event -- scripts/common.sh@364 -- # (( v = 0 )) 00:25:11.886 13:45:18 event -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:25:11.886 13:45:18 event -- scripts/common.sh@365 -- # decimal 1 00:25:11.886 13:45:18 event -- scripts/common.sh@353 -- # local d=1 00:25:11.886 13:45:18 event -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:25:11.886 13:45:18 event -- scripts/common.sh@355 -- # echo 1 00:25:11.886 13:45:18 event -- scripts/common.sh@365 -- # ver1[v]=1 00:25:11.886 13:45:18 event -- scripts/common.sh@366 -- # decimal 2 00:25:11.886 13:45:18 event -- scripts/common.sh@353 -- # local d=2 00:25:11.886 13:45:18 event -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:25:11.886 13:45:18 event -- scripts/common.sh@355 -- # echo 2 00:25:11.886 13:45:18 event -- scripts/common.sh@366 -- # ver2[v]=2 00:25:11.886 13:45:18 event -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:25:11.886 13:45:18 event -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:25:11.886 13:45:18 event -- scripts/common.sh@368 -- # return 0 00:25:11.886 13:45:18 event -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:25:11.886 13:45:18 event -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:25:11.886 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:11.886 --rc genhtml_branch_coverage=1 00:25:11.886 --rc genhtml_function_coverage=1 00:25:11.886 --rc genhtml_legend=1 00:25:11.886 --rc geninfo_all_blocks=1 00:25:11.886 --rc geninfo_unexecuted_blocks=1 00:25:11.886 00:25:11.886 ' 00:25:11.886 13:45:18 event -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:25:11.886 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:11.886 --rc genhtml_branch_coverage=1 00:25:11.886 --rc genhtml_function_coverage=1 00:25:11.886 --rc genhtml_legend=1 00:25:11.886 --rc geninfo_all_blocks=1 00:25:11.886 --rc geninfo_unexecuted_blocks=1 00:25:11.886 00:25:11.886 ' 00:25:11.886 13:45:18 event -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:25:11.886 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:11.886 --rc genhtml_branch_coverage=1 00:25:11.886 --rc genhtml_function_coverage=1 00:25:11.886 --rc genhtml_legend=1 00:25:11.886 --rc geninfo_all_blocks=1 00:25:11.886 --rc geninfo_unexecuted_blocks=1 00:25:11.886 00:25:11.886 ' 00:25:11.886 13:45:18 event -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:25:11.886 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:11.886 --rc genhtml_branch_coverage=1 00:25:11.886 --rc genhtml_function_coverage=1 00:25:11.886 --rc genhtml_legend=1 00:25:11.886 --rc geninfo_all_blocks=1 00:25:11.886 --rc geninfo_unexecuted_blocks=1 00:25:11.886 00:25:11.886 ' 00:25:11.886 13:45:18 event -- event/event.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:25:11.886 13:45:18 event -- bdev/nbd_common.sh@6 -- # set -e 00:25:11.886 13:45:18 event -- event/event.sh@45 -- # run_test event_perf /home/vagrant/spdk_repo/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:25:11.886 13:45:18 event -- common/autotest_common.sh@1105 -- # '[' 6 -le 1 ']' 00:25:11.886 13:45:18 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:25:11.886 13:45:18 event -- common/autotest_common.sh@10 -- # set +x 00:25:11.886 ************************************ 00:25:11.886 START TEST event_perf 00:25:11.886 ************************************ 00:25:11.886 13:45:18 event.event_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:25:11.886 Running I/O for 1 seconds...[2024-11-20 13:45:18.867750] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:25:11.886 [2024-11-20 13:45:18.867957] [ DPDK EAL parameters: event_perf --no-shconf -c 0xF --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57190 ] 00:25:12.144 [2024-11-20 13:45:19.026846] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:25:12.144 [2024-11-20 13:45:19.150268] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:25:12.144 [2024-11-20 13:45:19.150603] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:12.144 Running I/O for 1 seconds...[2024-11-20 13:45:19.150607] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:25:12.144 [2024-11-20 13:45:19.151159] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:25:13.518 00:25:13.518 lcore 0: 195948 00:25:13.518 lcore 1: 195950 00:25:13.518 lcore 2: 195952 00:25:13.518 lcore 3: 195953 00:25:13.518 done. 00:25:13.518 00:25:13.518 real 0m1.500s 00:25:13.518 user 0m4.274s 00:25:13.518 sys 0m0.100s 00:25:13.518 13:45:20 event.event_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:25:13.518 13:45:20 event.event_perf -- common/autotest_common.sh@10 -- # set +x 00:25:13.518 ************************************ 00:25:13.518 END TEST event_perf 00:25:13.518 ************************************ 00:25:13.518 13:45:20 event -- event/event.sh@46 -- # run_test event_reactor /home/vagrant/spdk_repo/spdk/test/event/reactor/reactor -t 1 00:25:13.518 13:45:20 event -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:25:13.518 13:45:20 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:25:13.518 13:45:20 event -- common/autotest_common.sh@10 -- # set +x 00:25:13.518 ************************************ 00:25:13.518 START TEST event_reactor 00:25:13.518 ************************************ 00:25:13.518 13:45:20 event.event_reactor -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/reactor/reactor -t 1 00:25:13.518 [2024-11-20 13:45:20.408497] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:25:13.518 [2024-11-20 13:45:20.408750] [ DPDK EAL parameters: reactor --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57230 ] 00:25:13.518 [2024-11-20 13:45:20.568822] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:13.776 [2024-11-20 13:45:20.683390] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:15.184 test_start 00:25:15.184 oneshot 00:25:15.184 tick 100 00:25:15.184 tick 100 00:25:15.184 tick 250 00:25:15.184 tick 100 00:25:15.184 tick 100 00:25:15.184 tick 100 00:25:15.184 tick 250 00:25:15.184 tick 500 00:25:15.184 tick 100 00:25:15.184 tick 100 00:25:15.184 tick 250 00:25:15.184 tick 100 00:25:15.185 tick 100 00:25:15.185 test_end 00:25:15.185 ************************************ 00:25:15.185 END TEST event_reactor 00:25:15.185 ************************************ 00:25:15.185 00:25:15.185 real 0m1.470s 00:25:15.185 user 0m1.294s 00:25:15.185 sys 0m0.068s 00:25:15.185 13:45:21 event.event_reactor -- common/autotest_common.sh@1130 -- # xtrace_disable 00:25:15.185 13:45:21 event.event_reactor -- common/autotest_common.sh@10 -- # set +x 00:25:15.185 13:45:21 event -- event/event.sh@47 -- # run_test event_reactor_perf /home/vagrant/spdk_repo/spdk/test/event/reactor_perf/reactor_perf -t 1 00:25:15.185 13:45:21 event -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:25:15.185 13:45:21 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:25:15.185 13:45:21 event -- common/autotest_common.sh@10 -- # set +x 00:25:15.185 ************************************ 00:25:15.185 START TEST event_reactor_perf 00:25:15.185 ************************************ 00:25:15.185 13:45:21 event.event_reactor_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/reactor_perf/reactor_perf -t 1 00:25:15.185 [2024-11-20 13:45:21.919199] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:25:15.185 [2024-11-20 13:45:21.919518] [ DPDK EAL parameters: reactor_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57266 ] 00:25:15.185 [2024-11-20 13:45:22.082610] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:15.185 [2024-11-20 13:45:22.195743] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:16.582 test_start 00:25:16.582 test_end 00:25:16.582 Performance: 298920 events per second 00:25:16.582 ************************************ 00:25:16.582 END TEST event_reactor_perf 00:25:16.582 ************************************ 00:25:16.582 00:25:16.582 real 0m1.477s 00:25:16.582 user 0m1.301s 00:25:16.582 sys 0m0.066s 00:25:16.582 13:45:23 event.event_reactor_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:25:16.582 13:45:23 event.event_reactor_perf -- common/autotest_common.sh@10 -- # set +x 00:25:16.582 13:45:23 event -- event/event.sh@49 -- # uname -s 00:25:16.582 13:45:23 event -- event/event.sh@49 -- # '[' Linux = Linux ']' 00:25:16.582 13:45:23 event -- event/event.sh@50 -- # run_test event_scheduler /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler.sh 00:25:16.582 13:45:23 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:25:16.582 13:45:23 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:25:16.582 13:45:23 event -- common/autotest_common.sh@10 -- # set +x 00:25:16.582 ************************************ 00:25:16.582 START TEST event_scheduler 00:25:16.582 ************************************ 00:25:16.582 13:45:23 event.event_scheduler -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler.sh 00:25:16.582 * Looking for test storage... 00:25:16.582 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event/scheduler 00:25:16.582 13:45:23 event.event_scheduler -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:25:16.583 13:45:23 event.event_scheduler -- common/autotest_common.sh@1693 -- # lcov --version 00:25:16.583 13:45:23 event.event_scheduler -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:25:16.583 13:45:23 event.event_scheduler -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@333 -- # local ver1 ver1_l 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@334 -- # local ver2 ver2_l 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@336 -- # IFS=.-: 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@336 -- # read -ra ver1 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@337 -- # IFS=.-: 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@337 -- # read -ra ver2 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@338 -- # local 'op=<' 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@340 -- # ver1_l=2 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@341 -- # ver2_l=1 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@344 -- # case "$op" in 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@345 -- # : 1 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@364 -- # (( v = 0 )) 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@365 -- # decimal 1 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@353 -- # local d=1 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@355 -- # echo 1 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@365 -- # ver1[v]=1 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@366 -- # decimal 2 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@353 -- # local d=2 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@355 -- # echo 2 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@366 -- # ver2[v]=2 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:25:16.583 13:45:23 event.event_scheduler -- scripts/common.sh@368 -- # return 0 00:25:16.583 13:45:23 event.event_scheduler -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:25:16.583 13:45:23 event.event_scheduler -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:25:16.583 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:16.583 --rc genhtml_branch_coverage=1 00:25:16.583 --rc genhtml_function_coverage=1 00:25:16.583 --rc genhtml_legend=1 00:25:16.583 --rc geninfo_all_blocks=1 00:25:16.583 --rc geninfo_unexecuted_blocks=1 00:25:16.583 00:25:16.583 ' 00:25:16.583 13:45:23 event.event_scheduler -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:25:16.583 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:16.583 --rc genhtml_branch_coverage=1 00:25:16.583 --rc genhtml_function_coverage=1 00:25:16.583 --rc genhtml_legend=1 00:25:16.583 --rc geninfo_all_blocks=1 00:25:16.583 --rc geninfo_unexecuted_blocks=1 00:25:16.583 00:25:16.583 ' 00:25:16.583 13:45:23 event.event_scheduler -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:25:16.583 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:16.583 --rc genhtml_branch_coverage=1 00:25:16.583 --rc genhtml_function_coverage=1 00:25:16.583 --rc genhtml_legend=1 00:25:16.583 --rc geninfo_all_blocks=1 00:25:16.583 --rc geninfo_unexecuted_blocks=1 00:25:16.583 00:25:16.583 ' 00:25:16.583 13:45:23 event.event_scheduler -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:25:16.583 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:16.583 --rc genhtml_branch_coverage=1 00:25:16.583 --rc genhtml_function_coverage=1 00:25:16.583 --rc genhtml_legend=1 00:25:16.583 --rc geninfo_all_blocks=1 00:25:16.583 --rc geninfo_unexecuted_blocks=1 00:25:16.583 00:25:16.583 ' 00:25:16.583 13:45:23 event.event_scheduler -- scheduler/scheduler.sh@29 -- # rpc=rpc_cmd 00:25:16.583 13:45:23 event.event_scheduler -- scheduler/scheduler.sh@35 -- # scheduler_pid=57337 00:25:16.583 13:45:23 event.event_scheduler -- scheduler/scheduler.sh@36 -- # trap 'killprocess $scheduler_pid; exit 1' SIGINT SIGTERM EXIT 00:25:16.583 13:45:23 event.event_scheduler -- scheduler/scheduler.sh@37 -- # waitforlisten 57337 00:25:16.583 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:16.583 13:45:23 event.event_scheduler -- common/autotest_common.sh@835 -- # '[' -z 57337 ']' 00:25:16.583 13:45:23 event.event_scheduler -- scheduler/scheduler.sh@34 -- # /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler -m 0xF -p 0x2 --wait-for-rpc -f 00:25:16.583 13:45:23 event.event_scheduler -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:16.583 13:45:23 event.event_scheduler -- common/autotest_common.sh@840 -- # local max_retries=100 00:25:16.583 13:45:23 event.event_scheduler -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:16.583 13:45:23 event.event_scheduler -- common/autotest_common.sh@844 -- # xtrace_disable 00:25:16.583 13:45:23 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:25:16.583 [2024-11-20 13:45:23.629049] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:25:16.583 [2024-11-20 13:45:23.629360] [ DPDK EAL parameters: scheduler --no-shconf -c 0xF --main-lcore=2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57337 ] 00:25:16.841 [2024-11-20 13:45:23.805184] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:25:17.100 [2024-11-20 13:45:23.950492] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:17.100 [2024-11-20 13:45:23.950864] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:25:17.100 [2024-11-20 13:45:23.950968] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:25:17.100 [2024-11-20 13:45:23.951005] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:25:17.665 13:45:24 event.event_scheduler -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:25:17.665 13:45:24 event.event_scheduler -- common/autotest_common.sh@868 -- # return 0 00:25:17.665 13:45:24 event.event_scheduler -- scheduler/scheduler.sh@39 -- # rpc_cmd framework_set_scheduler dynamic 00:25:17.665 13:45:24 event.event_scheduler -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:17.665 13:45:24 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:25:17.665 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:25:17.665 POWER: Cannot set governor of lcore 0 to userspace 00:25:17.665 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:25:17.665 POWER: Cannot set governor of lcore 0 to performance 00:25:17.665 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:25:17.665 POWER: Cannot set governor of lcore 0 to userspace 00:25:17.665 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:25:17.665 POWER: Cannot set governor of lcore 0 to userspace 00:25:17.665 GUEST_CHANNEL: Opening channel '/dev/virtio-ports/virtio.serial.port.poweragent.0' for lcore 0 00:25:17.665 GUEST_CHANNEL: Unable to connect to '/dev/virtio-ports/virtio.serial.port.poweragent.0' with error No such file or directory 00:25:17.665 POWER: Unable to set Power Management Environment for lcore 0 00:25:17.665 [2024-11-20 13:45:24.477801] dpdk_governor.c: 135:_init_core: *ERROR*: Failed to initialize on core0 00:25:17.665 [2024-11-20 13:45:24.477932] dpdk_governor.c: 196:_init: *ERROR*: Failed to initialize on core0 00:25:17.665 [2024-11-20 13:45:24.477961] scheduler_dynamic.c: 280:init: *NOTICE*: Unable to initialize dpdk governor 00:25:17.665 [2024-11-20 13:45:24.478094] scheduler_dynamic.c: 427:set_opts: *NOTICE*: Setting scheduler load limit to 20 00:25:17.665 [2024-11-20 13:45:24.478215] scheduler_dynamic.c: 429:set_opts: *NOTICE*: Setting scheduler core limit to 80 00:25:17.665 [2024-11-20 13:45:24.478265] scheduler_dynamic.c: 431:set_opts: *NOTICE*: Setting scheduler core busy to 95 00:25:17.665 13:45:24 event.event_scheduler -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:17.665 13:45:24 event.event_scheduler -- scheduler/scheduler.sh@40 -- # rpc_cmd framework_start_init 00:25:17.665 13:45:24 event.event_scheduler -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:17.665 13:45:24 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:25:17.924 [2024-11-20 13:45:24.726213] scheduler.c: 382:test_start: *NOTICE*: Scheduler test application started. 00:25:17.924 13:45:24 event.event_scheduler -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:17.924 13:45:24 event.event_scheduler -- scheduler/scheduler.sh@43 -- # run_test scheduler_create_thread scheduler_create_thread 00:25:17.924 13:45:24 event.event_scheduler -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:25:17.924 13:45:24 event.event_scheduler -- common/autotest_common.sh@1111 -- # xtrace_disable 00:25:17.924 13:45:24 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:25:17.924 ************************************ 00:25:17.924 START TEST scheduler_create_thread 00:25:17.924 ************************************ 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1129 -- # scheduler_create_thread 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@12 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x1 -a 100 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:25:17.924 2 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@13 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x2 -a 100 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:25:17.924 3 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@14 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x4 -a 100 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:25:17.924 4 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@15 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x8 -a 100 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:25:17.924 5 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@16 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x1 -a 0 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:25:17.924 6 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@17 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x2 -a 0 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:25:17.924 7 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@18 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x4 -a 0 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:25:17.924 8 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@19 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x8 -a 0 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:25:17.924 9 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@21 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n one_third_active -a 30 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:25:17.924 10 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n half_active -a 0 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # thread_id=11 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@23 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_set_active 11 50 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n deleted -a 100 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # thread_id=12 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@26 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_delete 12 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:17.924 13:45:24 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:25:18.491 ************************************ 00:25:18.491 END TEST scheduler_create_thread 00:25:18.491 ************************************ 00:25:18.491 13:45:25 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:18.491 00:25:18.491 real 0m0.593s 00:25:18.491 user 0m0.011s 00:25:18.491 sys 0m0.008s 00:25:18.491 13:45:25 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1130 -- # xtrace_disable 00:25:18.491 13:45:25 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:25:18.491 13:45:25 event.event_scheduler -- scheduler/scheduler.sh@45 -- # trap - SIGINT SIGTERM EXIT 00:25:18.491 13:45:25 event.event_scheduler -- scheduler/scheduler.sh@46 -- # killprocess 57337 00:25:18.491 13:45:25 event.event_scheduler -- common/autotest_common.sh@954 -- # '[' -z 57337 ']' 00:25:18.491 13:45:25 event.event_scheduler -- common/autotest_common.sh@958 -- # kill -0 57337 00:25:18.491 13:45:25 event.event_scheduler -- common/autotest_common.sh@959 -- # uname 00:25:18.491 13:45:25 event.event_scheduler -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:25:18.491 13:45:25 event.event_scheduler -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 57337 00:25:18.491 killing process with pid 57337 00:25:18.491 13:45:25 event.event_scheduler -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:25:18.491 13:45:25 event.event_scheduler -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:25:18.491 13:45:25 event.event_scheduler -- common/autotest_common.sh@972 -- # echo 'killing process with pid 57337' 00:25:18.491 13:45:25 event.event_scheduler -- common/autotest_common.sh@973 -- # kill 57337 00:25:18.491 13:45:25 event.event_scheduler -- common/autotest_common.sh@978 -- # wait 57337 00:25:19.056 [2024-11-20 13:45:25.812155] scheduler.c: 360:test_shutdown: *NOTICE*: Scheduler test application stopped. 00:25:19.621 00:25:19.621 real 0m3.201s 00:25:19.621 user 0m5.923s 00:25:19.621 sys 0m0.409s 00:25:19.621 ************************************ 00:25:19.621 END TEST event_scheduler 00:25:19.621 ************************************ 00:25:19.621 13:45:26 event.event_scheduler -- common/autotest_common.sh@1130 -- # xtrace_disable 00:25:19.621 13:45:26 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:25:19.621 13:45:26 event -- event/event.sh@51 -- # modprobe -n nbd 00:25:19.621 13:45:26 event -- event/event.sh@52 -- # run_test app_repeat app_repeat_test 00:25:19.621 13:45:26 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:25:19.621 13:45:26 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:25:19.621 13:45:26 event -- common/autotest_common.sh@10 -- # set +x 00:25:19.621 ************************************ 00:25:19.621 START TEST app_repeat 00:25:19.621 ************************************ 00:25:19.621 13:45:26 event.app_repeat -- common/autotest_common.sh@1129 -- # app_repeat_test 00:25:19.621 13:45:26 event.app_repeat -- event/event.sh@12 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:25:19.621 13:45:26 event.app_repeat -- event/event.sh@13 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:25:19.621 13:45:26 event.app_repeat -- event/event.sh@13 -- # local nbd_list 00:25:19.621 13:45:26 event.app_repeat -- event/event.sh@14 -- # bdev_list=('Malloc0' 'Malloc1') 00:25:19.621 13:45:26 event.app_repeat -- event/event.sh@14 -- # local bdev_list 00:25:19.621 13:45:26 event.app_repeat -- event/event.sh@15 -- # local repeat_times=4 00:25:19.621 13:45:26 event.app_repeat -- event/event.sh@17 -- # modprobe nbd 00:25:19.621 Process app_repeat pid: 57421 00:25:19.621 spdk_app_start Round 0 00:25:19.621 13:45:26 event.app_repeat -- event/event.sh@19 -- # repeat_pid=57421 00:25:19.621 13:45:26 event.app_repeat -- event/event.sh@20 -- # trap 'killprocess $repeat_pid; exit 1' SIGINT SIGTERM EXIT 00:25:19.621 13:45:26 event.app_repeat -- event/event.sh@21 -- # echo 'Process app_repeat pid: 57421' 00:25:19.621 13:45:26 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:25:19.621 13:45:26 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 0' 00:25:19.621 13:45:26 event.app_repeat -- event/event.sh@18 -- # /home/vagrant/spdk_repo/spdk/test/event/app_repeat/app_repeat -r /var/tmp/spdk-nbd.sock -m 0x3 -t 4 00:25:19.621 13:45:26 event.app_repeat -- event/event.sh@25 -- # waitforlisten 57421 /var/tmp/spdk-nbd.sock 00:25:19.621 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:25:19.621 13:45:26 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 57421 ']' 00:25:19.621 13:45:26 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:25:19.621 13:45:26 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:25:19.621 13:45:26 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:25:19.621 13:45:26 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:25:19.621 13:45:26 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:25:19.879 [2024-11-20 13:45:26.704451] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:25:19.879 [2024-11-20 13:45:26.704578] [ DPDK EAL parameters: app_repeat --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57421 ] 00:25:19.879 [2024-11-20 13:45:26.870343] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:25:20.136 [2024-11-20 13:45:26.989343] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:25:20.136 [2024-11-20 13:45:26.989535] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:20.701 13:45:27 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:25:20.701 13:45:27 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:25:20.701 13:45:27 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:25:20.959 Malloc0 00:25:20.959 13:45:27 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:25:21.216 Malloc1 00:25:21.216 13:45:28 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:25:21.216 13:45:28 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:25:21.216 13:45:28 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:25:21.216 13:45:28 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:25:21.216 13:45:28 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:25:21.216 13:45:28 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:25:21.216 13:45:28 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:25:21.216 13:45:28 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:25:21.216 13:45:28 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:25:21.216 13:45:28 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:25:21.216 13:45:28 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:25:21.216 13:45:28 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:25:21.216 13:45:28 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:25:21.216 13:45:28 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:25:21.216 13:45:28 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:25:21.216 13:45:28 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:25:21.473 /dev/nbd0 00:25:21.473 13:45:28 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:25:21.473 13:45:28 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:25:21.473 13:45:28 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:25:21.473 13:45:28 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:25:21.473 13:45:28 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:25:21.473 13:45:28 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:25:21.473 13:45:28 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:25:21.473 13:45:28 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:25:21.473 13:45:28 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:25:21.473 13:45:28 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:25:21.473 13:45:28 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:25:21.473 1+0 records in 00:25:21.473 1+0 records out 00:25:21.473 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00023193 s, 17.7 MB/s 00:25:21.473 13:45:28 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:25:21.473 13:45:28 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:25:21.473 13:45:28 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:25:21.473 13:45:28 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:25:21.473 13:45:28 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:25:21.473 13:45:28 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:25:21.473 13:45:28 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:25:21.473 13:45:28 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:25:21.743 /dev/nbd1 00:25:21.743 13:45:28 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:25:21.743 13:45:28 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:25:21.743 13:45:28 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:25:21.743 13:45:28 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:25:21.743 13:45:28 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:25:21.743 13:45:28 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:25:21.743 13:45:28 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:25:21.743 13:45:28 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:25:21.743 13:45:28 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:25:21.743 13:45:28 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:25:21.743 13:45:28 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:25:21.743 1+0 records in 00:25:21.743 1+0 records out 00:25:21.743 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000255299 s, 16.0 MB/s 00:25:21.743 13:45:28 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:25:21.743 13:45:28 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:25:21.743 13:45:28 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:25:21.743 13:45:28 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:25:21.743 13:45:28 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:25:21.743 13:45:28 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:25:21.743 13:45:28 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:25:21.743 13:45:28 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:25:21.743 13:45:28 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:25:21.743 13:45:28 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:25:21.743 13:45:28 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:25:21.743 { 00:25:21.743 "nbd_device": "/dev/nbd0", 00:25:21.743 "bdev_name": "Malloc0" 00:25:21.743 }, 00:25:21.743 { 00:25:21.743 "nbd_device": "/dev/nbd1", 00:25:21.743 "bdev_name": "Malloc1" 00:25:21.743 } 00:25:21.743 ]' 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:25:22.003 { 00:25:22.003 "nbd_device": "/dev/nbd0", 00:25:22.003 "bdev_name": "Malloc0" 00:25:22.003 }, 00:25:22.003 { 00:25:22.003 "nbd_device": "/dev/nbd1", 00:25:22.003 "bdev_name": "Malloc1" 00:25:22.003 } 00:25:22.003 ]' 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:25:22.003 /dev/nbd1' 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:25:22.003 /dev/nbd1' 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:25:22.003 256+0 records in 00:25:22.003 256+0 records out 00:25:22.003 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00684013 s, 153 MB/s 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:25:22.003 256+0 records in 00:25:22.003 256+0 records out 00:25:22.003 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0209396 s, 50.1 MB/s 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:25:22.003 256+0 records in 00:25:22.003 256+0 records out 00:25:22.003 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0201565 s, 52.0 MB/s 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:25:22.003 13:45:28 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:25:22.272 13:45:29 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:25:22.272 13:45:29 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:25:22.272 13:45:29 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:25:22.272 13:45:29 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:25:22.272 13:45:29 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:25:22.272 13:45:29 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:25:22.272 13:45:29 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:25:22.272 13:45:29 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:25:22.272 13:45:29 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:25:22.272 13:45:29 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:25:22.272 13:45:29 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:25:22.272 13:45:29 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:25:22.272 13:45:29 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:25:22.272 13:45:29 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:25:22.272 13:45:29 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:25:22.272 13:45:29 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:25:22.272 13:45:29 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:25:22.272 13:45:29 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:25:22.528 13:45:29 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:25:22.528 13:45:29 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:25:22.529 13:45:29 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:25:22.529 13:45:29 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:25:22.529 13:45:29 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:25:22.529 13:45:29 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:25:22.529 13:45:29 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:25:22.529 13:45:29 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:25:22.529 13:45:29 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:25:22.529 13:45:29 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:25:22.529 13:45:29 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:25:22.529 13:45:29 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:25:22.529 13:45:29 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:25:22.529 13:45:29 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:25:22.529 13:45:29 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:25:22.529 13:45:29 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:25:23.092 13:45:29 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:25:23.657 [2024-11-20 13:45:30.663268] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:25:23.915 [2024-11-20 13:45:30.773129] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:25:23.915 [2024-11-20 13:45:30.773436] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:23.915 [2024-11-20 13:45:30.905590] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:25:23.915 [2024-11-20 13:45:30.905688] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:25:26.441 spdk_app_start Round 1 00:25:26.441 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:25:26.441 13:45:32 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:25:26.441 13:45:32 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 1' 00:25:26.441 13:45:32 event.app_repeat -- event/event.sh@25 -- # waitforlisten 57421 /var/tmp/spdk-nbd.sock 00:25:26.441 13:45:32 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 57421 ']' 00:25:26.441 13:45:32 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:25:26.441 13:45:32 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:25:26.441 13:45:32 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:25:26.441 13:45:32 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:25:26.441 13:45:32 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:25:26.441 13:45:33 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:25:26.441 13:45:33 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:25:26.441 13:45:33 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:25:26.441 Malloc0 00:25:26.441 13:45:33 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:25:26.700 Malloc1 00:25:26.700 13:45:33 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:25:26.700 13:45:33 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:25:26.700 13:45:33 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:25:26.700 13:45:33 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:25:26.700 13:45:33 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:25:26.700 13:45:33 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:25:26.700 13:45:33 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:25:26.700 13:45:33 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:25:26.700 13:45:33 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:25:26.700 13:45:33 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:25:26.700 13:45:33 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:25:26.700 13:45:33 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:25:26.700 13:45:33 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:25:26.700 13:45:33 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:25:26.700 13:45:33 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:25:26.700 13:45:33 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:25:26.958 /dev/nbd0 00:25:26.958 13:45:33 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:25:26.958 13:45:33 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:25:26.958 13:45:33 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:25:26.958 13:45:33 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:25:26.958 13:45:33 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:25:26.958 13:45:33 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:25:26.958 13:45:33 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:25:26.958 13:45:33 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:25:26.958 13:45:33 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:25:26.958 13:45:33 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:25:26.958 13:45:33 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:25:26.958 1+0 records in 00:25:26.958 1+0 records out 00:25:26.958 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00065775 s, 6.2 MB/s 00:25:26.958 13:45:33 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:25:26.958 13:45:33 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:25:26.958 13:45:33 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:25:26.958 13:45:33 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:25:26.958 13:45:33 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:25:26.958 13:45:33 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:25:26.958 13:45:33 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:25:26.958 13:45:33 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:25:27.215 /dev/nbd1 00:25:27.216 13:45:34 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:25:27.216 13:45:34 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:25:27.216 13:45:34 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:25:27.216 13:45:34 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:25:27.216 13:45:34 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:25:27.216 13:45:34 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:25:27.216 13:45:34 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:25:27.216 13:45:34 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:25:27.216 13:45:34 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:25:27.216 13:45:34 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:25:27.216 13:45:34 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:25:27.216 1+0 records in 00:25:27.216 1+0 records out 00:25:27.216 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000225893 s, 18.1 MB/s 00:25:27.216 13:45:34 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:25:27.216 13:45:34 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:25:27.216 13:45:34 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:25:27.216 13:45:34 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:25:27.216 13:45:34 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:25:27.216 13:45:34 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:25:27.216 13:45:34 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:25:27.216 13:45:34 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:25:27.216 13:45:34 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:25:27.216 13:45:34 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:25:27.216 13:45:34 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:25:27.216 { 00:25:27.216 "nbd_device": "/dev/nbd0", 00:25:27.216 "bdev_name": "Malloc0" 00:25:27.216 }, 00:25:27.216 { 00:25:27.216 "nbd_device": "/dev/nbd1", 00:25:27.216 "bdev_name": "Malloc1" 00:25:27.216 } 00:25:27.216 ]' 00:25:27.216 13:45:34 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:25:27.216 13:45:34 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:25:27.216 { 00:25:27.216 "nbd_device": "/dev/nbd0", 00:25:27.216 "bdev_name": "Malloc0" 00:25:27.216 }, 00:25:27.216 { 00:25:27.216 "nbd_device": "/dev/nbd1", 00:25:27.216 "bdev_name": "Malloc1" 00:25:27.216 } 00:25:27.216 ]' 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:25:27.474 /dev/nbd1' 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:25:27.474 /dev/nbd1' 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:25:27.474 256+0 records in 00:25:27.474 256+0 records out 00:25:27.474 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00731524 s, 143 MB/s 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:25:27.474 256+0 records in 00:25:27.474 256+0 records out 00:25:27.474 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0185641 s, 56.5 MB/s 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:25:27.474 256+0 records in 00:25:27.474 256+0 records out 00:25:27.474 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0225561 s, 46.5 MB/s 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:25:27.474 13:45:34 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:25:27.733 13:45:34 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:25:27.733 13:45:34 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:25:27.733 13:45:34 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:25:27.733 13:45:34 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:25:27.733 13:45:34 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:25:27.733 13:45:34 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:25:27.733 13:45:34 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:25:27.733 13:45:34 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:25:27.733 13:45:34 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:25:27.733 13:45:34 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:25:27.733 13:45:34 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:25:27.733 13:45:34 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:25:27.733 13:45:34 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:25:27.733 13:45:34 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:25:27.733 13:45:34 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:25:27.733 13:45:34 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:25:27.733 13:45:34 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:25:27.733 13:45:34 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:25:27.733 13:45:34 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:25:27.733 13:45:34 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:25:27.733 13:45:34 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:25:27.991 13:45:34 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:25:27.991 13:45:34 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:25:27.991 13:45:34 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:25:27.991 13:45:35 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:25:27.991 13:45:35 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:25:27.991 13:45:35 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:25:27.991 13:45:35 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:25:27.991 13:45:35 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:25:27.991 13:45:35 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:25:27.991 13:45:35 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:25:27.991 13:45:35 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:25:27.991 13:45:35 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:25:27.991 13:45:35 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:25:28.557 13:45:35 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:25:29.123 [2024-11-20 13:45:35.975257] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:25:29.123 [2024-11-20 13:45:36.072928] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:29.123 [2024-11-20 13:45:36.072942] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:25:29.381 [2024-11-20 13:45:36.188405] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:25:29.381 [2024-11-20 13:45:36.188491] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:25:31.340 spdk_app_start Round 2 00:25:31.340 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:25:31.340 13:45:38 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:25:31.340 13:45:38 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 2' 00:25:31.340 13:45:38 event.app_repeat -- event/event.sh@25 -- # waitforlisten 57421 /var/tmp/spdk-nbd.sock 00:25:31.340 13:45:38 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 57421 ']' 00:25:31.340 13:45:38 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:25:31.340 13:45:38 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:25:31.340 13:45:38 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:25:31.340 13:45:38 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:25:31.340 13:45:38 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:25:31.598 13:45:38 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:25:31.598 13:45:38 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:25:31.598 13:45:38 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:25:31.856 Malloc0 00:25:31.856 13:45:38 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:25:32.113 Malloc1 00:25:32.113 13:45:39 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:25:32.113 13:45:39 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:25:32.113 13:45:39 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:25:32.114 13:45:39 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:25:32.114 13:45:39 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:25:32.114 13:45:39 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:25:32.114 13:45:39 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:25:32.114 13:45:39 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:25:32.114 13:45:39 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:25:32.114 13:45:39 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:25:32.114 13:45:39 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:25:32.114 13:45:39 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:25:32.114 13:45:39 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:25:32.114 13:45:39 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:25:32.114 13:45:39 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:25:32.114 13:45:39 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:25:32.371 /dev/nbd0 00:25:32.371 13:45:39 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:25:32.371 13:45:39 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:25:32.371 13:45:39 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:25:32.371 13:45:39 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:25:32.371 13:45:39 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:25:32.371 13:45:39 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:25:32.371 13:45:39 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:25:32.371 13:45:39 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:25:32.371 13:45:39 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:25:32.371 13:45:39 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:25:32.371 13:45:39 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:25:32.371 1+0 records in 00:25:32.371 1+0 records out 00:25:32.371 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000169524 s, 24.2 MB/s 00:25:32.371 13:45:39 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:25:32.371 13:45:39 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:25:32.371 13:45:39 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:25:32.371 13:45:39 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:25:32.371 13:45:39 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:25:32.371 13:45:39 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:25:32.371 13:45:39 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:25:32.371 13:45:39 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:25:32.629 /dev/nbd1 00:25:32.629 13:45:39 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:25:32.629 13:45:39 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:25:32.629 13:45:39 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:25:32.629 13:45:39 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:25:32.629 13:45:39 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:25:32.629 13:45:39 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:25:32.629 13:45:39 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:25:32.629 13:45:39 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:25:32.629 13:45:39 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:25:32.629 13:45:39 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:25:32.629 13:45:39 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:25:32.629 1+0 records in 00:25:32.629 1+0 records out 00:25:32.629 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000203417 s, 20.1 MB/s 00:25:32.629 13:45:39 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:25:32.629 13:45:39 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:25:32.629 13:45:39 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:25:32.629 13:45:39 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:25:32.629 13:45:39 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:25:32.629 13:45:39 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:25:32.629 13:45:39 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:25:32.629 13:45:39 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:25:32.629 13:45:39 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:25:32.629 13:45:39 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:25:32.969 13:45:39 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:25:32.969 { 00:25:32.969 "nbd_device": "/dev/nbd0", 00:25:32.969 "bdev_name": "Malloc0" 00:25:32.969 }, 00:25:32.969 { 00:25:32.969 "nbd_device": "/dev/nbd1", 00:25:32.969 "bdev_name": "Malloc1" 00:25:32.969 } 00:25:32.969 ]' 00:25:32.969 13:45:39 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:25:32.969 { 00:25:32.969 "nbd_device": "/dev/nbd0", 00:25:32.969 "bdev_name": "Malloc0" 00:25:32.969 }, 00:25:32.969 { 00:25:32.969 "nbd_device": "/dev/nbd1", 00:25:32.969 "bdev_name": "Malloc1" 00:25:32.969 } 00:25:32.969 ]' 00:25:32.969 13:45:39 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:25:32.969 13:45:39 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:25:32.969 /dev/nbd1' 00:25:32.969 13:45:39 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:25:32.969 /dev/nbd1' 00:25:32.969 13:45:39 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:25:32.969 13:45:39 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:25:32.969 13:45:39 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:25:32.969 13:45:39 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:25:32.969 13:45:39 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:25:32.969 13:45:39 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:25:32.969 13:45:39 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:25:32.969 13:45:39 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:25:32.969 13:45:39 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:25:32.969 13:45:39 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:25:32.969 13:45:39 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:25:32.969 13:45:39 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:25:32.969 256+0 records in 00:25:32.969 256+0 records out 00:25:32.970 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00681515 s, 154 MB/s 00:25:32.970 13:45:39 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:25:32.970 13:45:39 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:25:32.970 256+0 records in 00:25:32.970 256+0 records out 00:25:32.970 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0165298 s, 63.4 MB/s 00:25:32.970 13:45:39 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:25:32.970 13:45:39 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:25:32.970 256+0 records in 00:25:32.970 256+0 records out 00:25:32.970 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0207566 s, 50.5 MB/s 00:25:32.970 13:45:39 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:25:32.970 13:45:39 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:25:32.970 13:45:39 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:25:32.970 13:45:39 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:25:32.970 13:45:39 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:25:32.970 13:45:39 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:25:32.970 13:45:39 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:25:32.970 13:45:39 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:25:32.970 13:45:39 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:25:32.970 13:45:39 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:25:32.970 13:45:39 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:25:32.970 13:45:39 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:25:32.970 13:45:39 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:25:32.970 13:45:39 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:25:32.970 13:45:39 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:25:32.970 13:45:39 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:25:32.970 13:45:39 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:25:32.970 13:45:39 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:25:32.970 13:45:39 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:25:33.228 13:45:40 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:25:33.228 13:45:40 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:25:33.228 13:45:40 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:25:33.228 13:45:40 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:25:33.228 13:45:40 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:25:33.228 13:45:40 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:25:33.228 13:45:40 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:25:33.228 13:45:40 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:25:33.228 13:45:40 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:25:33.228 13:45:40 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:25:33.486 13:45:40 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:25:33.486 13:45:40 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:25:33.486 13:45:40 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:25:33.486 13:45:40 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:25:33.486 13:45:40 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:25:33.486 13:45:40 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:25:33.486 13:45:40 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:25:33.486 13:45:40 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:25:33.486 13:45:40 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:25:33.486 13:45:40 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:25:33.486 13:45:40 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:25:33.486 13:45:40 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:25:33.486 13:45:40 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:25:33.486 13:45:40 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:25:33.745 13:45:40 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:25:33.745 13:45:40 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:25:33.745 13:45:40 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:25:33.745 13:45:40 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:25:33.745 13:45:40 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:25:33.745 13:45:40 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:25:33.745 13:45:40 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:25:33.745 13:45:40 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:25:33.745 13:45:40 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:25:33.745 13:45:40 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:25:34.003 13:45:40 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:25:34.568 [2024-11-20 13:45:41.481889] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:25:34.569 [2024-11-20 13:45:41.583478] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:25:34.569 [2024-11-20 13:45:41.583488] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:34.827 [2024-11-20 13:45:41.697451] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:25:34.827 [2024-11-20 13:45:41.697542] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:25:37.379 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:25:37.379 13:45:43 event.app_repeat -- event/event.sh@38 -- # waitforlisten 57421 /var/tmp/spdk-nbd.sock 00:25:37.379 13:45:43 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 57421 ']' 00:25:37.379 13:45:43 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:25:37.379 13:45:43 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:25:37.379 13:45:43 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:25:37.379 13:45:43 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:25:37.379 13:45:43 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:25:37.379 13:45:44 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:25:37.379 13:45:44 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:25:37.379 13:45:44 event.app_repeat -- event/event.sh@39 -- # killprocess 57421 00:25:37.379 13:45:44 event.app_repeat -- common/autotest_common.sh@954 -- # '[' -z 57421 ']' 00:25:37.379 13:45:44 event.app_repeat -- common/autotest_common.sh@958 -- # kill -0 57421 00:25:37.379 13:45:44 event.app_repeat -- common/autotest_common.sh@959 -- # uname 00:25:37.379 13:45:44 event.app_repeat -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:25:37.379 13:45:44 event.app_repeat -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 57421 00:25:37.379 killing process with pid 57421 00:25:37.379 13:45:44 event.app_repeat -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:25:37.379 13:45:44 event.app_repeat -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:25:37.379 13:45:44 event.app_repeat -- common/autotest_common.sh@972 -- # echo 'killing process with pid 57421' 00:25:37.379 13:45:44 event.app_repeat -- common/autotest_common.sh@973 -- # kill 57421 00:25:37.379 13:45:44 event.app_repeat -- common/autotest_common.sh@978 -- # wait 57421 00:25:37.636 spdk_app_start is called in Round 0. 00:25:37.636 Shutdown signal received, stop current app iteration 00:25:37.636 Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 reinitialization... 00:25:37.636 spdk_app_start is called in Round 1. 00:25:37.636 Shutdown signal received, stop current app iteration 00:25:37.636 Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 reinitialization... 00:25:37.636 spdk_app_start is called in Round 2. 00:25:37.636 Shutdown signal received, stop current app iteration 00:25:37.636 Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 reinitialization... 00:25:37.636 spdk_app_start is called in Round 3. 00:25:37.637 Shutdown signal received, stop current app iteration 00:25:37.637 13:45:44 event.app_repeat -- event/event.sh@40 -- # trap - SIGINT SIGTERM EXIT 00:25:37.637 13:45:44 event.app_repeat -- event/event.sh@42 -- # return 0 00:25:37.637 00:25:37.637 real 0m18.030s 00:25:37.637 user 0m39.310s 00:25:37.637 sys 0m2.220s 00:25:37.637 ************************************ 00:25:37.637 END TEST app_repeat 00:25:37.637 ************************************ 00:25:37.637 13:45:44 event.app_repeat -- common/autotest_common.sh@1130 -- # xtrace_disable 00:25:37.637 13:45:44 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:25:37.895 13:45:44 event -- event/event.sh@54 -- # (( SPDK_TEST_CRYPTO == 0 )) 00:25:37.895 13:45:44 event -- event/event.sh@55 -- # run_test cpu_locks /home/vagrant/spdk_repo/spdk/test/event/cpu_locks.sh 00:25:37.895 13:45:44 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:25:37.895 13:45:44 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:25:37.895 13:45:44 event -- common/autotest_common.sh@10 -- # set +x 00:25:37.895 ************************************ 00:25:37.895 START TEST cpu_locks 00:25:37.895 ************************************ 00:25:37.895 13:45:44 event.cpu_locks -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/cpu_locks.sh 00:25:37.895 * Looking for test storage... 00:25:37.895 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event 00:25:37.895 13:45:44 event.cpu_locks -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:25:37.895 13:45:44 event.cpu_locks -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:25:37.895 13:45:44 event.cpu_locks -- common/autotest_common.sh@1693 -- # lcov --version 00:25:37.895 13:45:44 event.cpu_locks -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@333 -- # local ver1 ver1_l 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@334 -- # local ver2 ver2_l 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@336 -- # IFS=.-: 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@336 -- # read -ra ver1 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@337 -- # IFS=.-: 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@337 -- # read -ra ver2 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@338 -- # local 'op=<' 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@340 -- # ver1_l=2 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@341 -- # ver2_l=1 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@344 -- # case "$op" in 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@345 -- # : 1 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@364 -- # (( v = 0 )) 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@365 -- # decimal 1 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@353 -- # local d=1 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@355 -- # echo 1 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@365 -- # ver1[v]=1 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@366 -- # decimal 2 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@353 -- # local d=2 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@355 -- # echo 2 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@366 -- # ver2[v]=2 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:25:37.895 13:45:44 event.cpu_locks -- scripts/common.sh@368 -- # return 0 00:25:37.895 13:45:44 event.cpu_locks -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:25:37.895 13:45:44 event.cpu_locks -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:25:37.895 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:37.895 --rc genhtml_branch_coverage=1 00:25:37.895 --rc genhtml_function_coverage=1 00:25:37.895 --rc genhtml_legend=1 00:25:37.895 --rc geninfo_all_blocks=1 00:25:37.895 --rc geninfo_unexecuted_blocks=1 00:25:37.895 00:25:37.895 ' 00:25:37.895 13:45:44 event.cpu_locks -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:25:37.895 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:37.895 --rc genhtml_branch_coverage=1 00:25:37.895 --rc genhtml_function_coverage=1 00:25:37.895 --rc genhtml_legend=1 00:25:37.895 --rc geninfo_all_blocks=1 00:25:37.895 --rc geninfo_unexecuted_blocks=1 00:25:37.895 00:25:37.895 ' 00:25:37.895 13:45:44 event.cpu_locks -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:25:37.895 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:37.895 --rc genhtml_branch_coverage=1 00:25:37.895 --rc genhtml_function_coverage=1 00:25:37.895 --rc genhtml_legend=1 00:25:37.895 --rc geninfo_all_blocks=1 00:25:37.895 --rc geninfo_unexecuted_blocks=1 00:25:37.895 00:25:37.895 ' 00:25:37.895 13:45:44 event.cpu_locks -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:25:37.895 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:37.895 --rc genhtml_branch_coverage=1 00:25:37.895 --rc genhtml_function_coverage=1 00:25:37.895 --rc genhtml_legend=1 00:25:37.895 --rc geninfo_all_blocks=1 00:25:37.895 --rc geninfo_unexecuted_blocks=1 00:25:37.895 00:25:37.895 ' 00:25:37.895 13:45:44 event.cpu_locks -- event/cpu_locks.sh@11 -- # rpc_sock1=/var/tmp/spdk.sock 00:25:37.895 13:45:44 event.cpu_locks -- event/cpu_locks.sh@12 -- # rpc_sock2=/var/tmp/spdk2.sock 00:25:37.895 13:45:44 event.cpu_locks -- event/cpu_locks.sh@164 -- # trap cleanup EXIT SIGTERM SIGINT 00:25:37.895 13:45:44 event.cpu_locks -- event/cpu_locks.sh@166 -- # run_test default_locks default_locks 00:25:37.895 13:45:44 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:25:37.895 13:45:44 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:25:37.895 13:45:44 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:25:37.895 ************************************ 00:25:37.895 START TEST default_locks 00:25:37.895 ************************************ 00:25:37.895 13:45:44 event.cpu_locks.default_locks -- common/autotest_common.sh@1129 -- # default_locks 00:25:37.895 13:45:44 event.cpu_locks.default_locks -- event/cpu_locks.sh@46 -- # spdk_tgt_pid=57857 00:25:37.895 13:45:44 event.cpu_locks.default_locks -- event/cpu_locks.sh@47 -- # waitforlisten 57857 00:25:37.895 13:45:44 event.cpu_locks.default_locks -- event/cpu_locks.sh@45 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:25:37.895 13:45:44 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # '[' -z 57857 ']' 00:25:37.895 13:45:44 event.cpu_locks.default_locks -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:37.895 13:45:44 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # local max_retries=100 00:25:37.896 13:45:44 event.cpu_locks.default_locks -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:37.896 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:37.896 13:45:44 event.cpu_locks.default_locks -- common/autotest_common.sh@844 -- # xtrace_disable 00:25:37.896 13:45:44 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:25:38.154 [2024-11-20 13:45:44.955231] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:25:38.154 [2024-11-20 13:45:44.955507] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57857 ] 00:25:38.154 [2024-11-20 13:45:45.109588] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:38.413 [2024-11-20 13:45:45.215179] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:38.979 13:45:45 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:25:38.979 13:45:45 event.cpu_locks.default_locks -- common/autotest_common.sh@868 -- # return 0 00:25:38.979 13:45:45 event.cpu_locks.default_locks -- event/cpu_locks.sh@49 -- # locks_exist 57857 00:25:38.979 13:45:45 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # lslocks -p 57857 00:25:38.979 13:45:45 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:25:38.979 13:45:45 event.cpu_locks.default_locks -- event/cpu_locks.sh@50 -- # killprocess 57857 00:25:38.979 13:45:45 event.cpu_locks.default_locks -- common/autotest_common.sh@954 -- # '[' -z 57857 ']' 00:25:38.979 13:45:45 event.cpu_locks.default_locks -- common/autotest_common.sh@958 -- # kill -0 57857 00:25:38.979 13:45:45 event.cpu_locks.default_locks -- common/autotest_common.sh@959 -- # uname 00:25:38.979 13:45:45 event.cpu_locks.default_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:25:38.979 13:45:45 event.cpu_locks.default_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 57857 00:25:38.979 killing process with pid 57857 00:25:38.979 13:45:45 event.cpu_locks.default_locks -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:25:38.979 13:45:45 event.cpu_locks.default_locks -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:25:38.979 13:45:45 event.cpu_locks.default_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 57857' 00:25:38.979 13:45:45 event.cpu_locks.default_locks -- common/autotest_common.sh@973 -- # kill 57857 00:25:38.979 13:45:45 event.cpu_locks.default_locks -- common/autotest_common.sh@978 -- # wait 57857 00:25:40.353 13:45:47 event.cpu_locks.default_locks -- event/cpu_locks.sh@52 -- # NOT waitforlisten 57857 00:25:40.353 13:45:47 event.cpu_locks.default_locks -- common/autotest_common.sh@652 -- # local es=0 00:25:40.353 13:45:47 event.cpu_locks.default_locks -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 57857 00:25:40.353 13:45:47 event.cpu_locks.default_locks -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:25:40.353 13:45:47 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:25:40.353 13:45:47 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:25:40.353 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:40.353 ERROR: process (pid: 57857) is no longer running 00:25:40.353 ************************************ 00:25:40.353 END TEST default_locks 00:25:40.353 ************************************ 00:25:40.353 13:45:47 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:25:40.353 13:45:47 event.cpu_locks.default_locks -- common/autotest_common.sh@655 -- # waitforlisten 57857 00:25:40.353 13:45:47 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # '[' -z 57857 ']' 00:25:40.353 13:45:47 event.cpu_locks.default_locks -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:40.353 13:45:47 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # local max_retries=100 00:25:40.353 13:45:47 event.cpu_locks.default_locks -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:40.353 13:45:47 event.cpu_locks.default_locks -- common/autotest_common.sh@844 -- # xtrace_disable 00:25:40.353 13:45:47 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:25:40.353 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 850: kill: (57857) - No such process 00:25:40.353 13:45:47 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:25:40.353 13:45:47 event.cpu_locks.default_locks -- common/autotest_common.sh@868 -- # return 1 00:25:40.353 13:45:47 event.cpu_locks.default_locks -- common/autotest_common.sh@655 -- # es=1 00:25:40.353 13:45:47 event.cpu_locks.default_locks -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:25:40.353 13:45:47 event.cpu_locks.default_locks -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:25:40.353 13:45:47 event.cpu_locks.default_locks -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:25:40.353 13:45:47 event.cpu_locks.default_locks -- event/cpu_locks.sh@54 -- # no_locks 00:25:40.353 13:45:47 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # lock_files=() 00:25:40.353 13:45:47 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # local lock_files 00:25:40.353 13:45:47 event.cpu_locks.default_locks -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:25:40.353 00:25:40.353 real 0m2.371s 00:25:40.353 user 0m2.334s 00:25:40.353 sys 0m0.469s 00:25:40.353 13:45:47 event.cpu_locks.default_locks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:25:40.353 13:45:47 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:25:40.353 13:45:47 event.cpu_locks -- event/cpu_locks.sh@167 -- # run_test default_locks_via_rpc default_locks_via_rpc 00:25:40.353 13:45:47 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:25:40.353 13:45:47 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:25:40.353 13:45:47 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:25:40.353 ************************************ 00:25:40.353 START TEST default_locks_via_rpc 00:25:40.353 ************************************ 00:25:40.353 13:45:47 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1129 -- # default_locks_via_rpc 00:25:40.353 13:45:47 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@62 -- # spdk_tgt_pid=57910 00:25:40.353 13:45:47 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@63 -- # waitforlisten 57910 00:25:40.353 13:45:47 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 57910 ']' 00:25:40.353 13:45:47 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:40.353 13:45:47 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:25:40.353 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:40.353 13:45:47 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:40.353 13:45:47 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:25:40.353 13:45:47 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:25:40.353 13:45:47 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@61 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:25:40.353 [2024-11-20 13:45:47.402024] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:25:40.353 [2024-11-20 13:45:47.402201] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57910 ] 00:25:40.612 [2024-11-20 13:45:47.576517] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:40.869 [2024-11-20 13:45:47.692102] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:41.437 13:45:48 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:25:41.437 13:45:48 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:25:41.437 13:45:48 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@65 -- # rpc_cmd framework_disable_cpumask_locks 00:25:41.437 13:45:48 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:41.437 13:45:48 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:25:41.437 13:45:48 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:41.437 13:45:48 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@67 -- # no_locks 00:25:41.437 13:45:48 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # lock_files=() 00:25:41.437 13:45:48 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # local lock_files 00:25:41.437 13:45:48 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:25:41.437 13:45:48 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@69 -- # rpc_cmd framework_enable_cpumask_locks 00:25:41.437 13:45:48 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:41.437 13:45:48 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:25:41.437 13:45:48 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:41.437 13:45:48 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@71 -- # locks_exist 57910 00:25:41.437 13:45:48 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:25:41.437 13:45:48 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # lslocks -p 57910 00:25:41.695 13:45:48 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@73 -- # killprocess 57910 00:25:41.696 13:45:48 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@954 -- # '[' -z 57910 ']' 00:25:41.696 13:45:48 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@958 -- # kill -0 57910 00:25:41.696 13:45:48 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@959 -- # uname 00:25:41.696 13:45:48 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:25:41.696 13:45:48 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 57910 00:25:41.696 killing process with pid 57910 00:25:41.696 13:45:48 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:25:41.696 13:45:48 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:25:41.696 13:45:48 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 57910' 00:25:41.696 13:45:48 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@973 -- # kill 57910 00:25:41.696 13:45:48 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@978 -- # wait 57910 00:25:43.597 00:25:43.597 real 0m2.927s 00:25:43.597 user 0m2.863s 00:25:43.597 sys 0m0.528s 00:25:43.597 13:45:50 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:25:43.597 ************************************ 00:25:43.597 END TEST default_locks_via_rpc 00:25:43.597 ************************************ 00:25:43.597 13:45:50 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:25:43.597 13:45:50 event.cpu_locks -- event/cpu_locks.sh@168 -- # run_test non_locking_app_on_locked_coremask non_locking_app_on_locked_coremask 00:25:43.597 13:45:50 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:25:43.597 13:45:50 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:25:43.597 13:45:50 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:25:43.597 ************************************ 00:25:43.597 START TEST non_locking_app_on_locked_coremask 00:25:43.597 ************************************ 00:25:43.597 13:45:50 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1129 -- # non_locking_app_on_locked_coremask 00:25:43.597 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:43.597 13:45:50 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@80 -- # spdk_tgt_pid=57973 00:25:43.597 13:45:50 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@81 -- # waitforlisten 57973 /var/tmp/spdk.sock 00:25:43.597 13:45:50 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 57973 ']' 00:25:43.597 13:45:50 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:43.597 13:45:50 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@79 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:25:43.597 13:45:50 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:25:43.597 13:45:50 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:43.597 13:45:50 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:25:43.597 13:45:50 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:25:43.597 [2024-11-20 13:45:50.337512] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:25:43.597 [2024-11-20 13:45:50.337639] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57973 ] 00:25:43.597 [2024-11-20 13:45:50.498142] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:43.597 [2024-11-20 13:45:50.614269] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:44.530 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:25:44.530 13:45:51 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:25:44.530 13:45:51 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:25:44.530 13:45:51 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@84 -- # spdk_tgt_pid2=57989 00:25:44.530 13:45:51 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@85 -- # waitforlisten 57989 /var/tmp/spdk2.sock 00:25:44.530 13:45:51 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 57989 ']' 00:25:44.530 13:45:51 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:25:44.530 13:45:51 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:25:44.530 13:45:51 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:25:44.530 13:45:51 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:25:44.530 13:45:51 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:25:44.530 13:45:51 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@83 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks -r /var/tmp/spdk2.sock 00:25:44.530 [2024-11-20 13:45:51.350729] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:25:44.530 [2024-11-20 13:45:51.350850] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57989 ] 00:25:44.530 [2024-11-20 13:45:51.525396] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:25:44.530 [2024-11-20 13:45:51.525474] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:44.788 [2024-11-20 13:45:51.756823] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:46.160 13:45:52 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:25:46.160 13:45:52 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:25:46.160 13:45:52 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@87 -- # locks_exist 57973 00:25:46.160 13:45:52 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 57973 00:25:46.160 13:45:52 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:25:46.160 13:45:53 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@89 -- # killprocess 57973 00:25:46.160 13:45:53 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 57973 ']' 00:25:46.160 13:45:53 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 57973 00:25:46.160 13:45:53 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:25:46.160 13:45:53 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:25:46.160 13:45:53 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 57973 00:25:46.418 killing process with pid 57973 00:25:46.418 13:45:53 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:25:46.418 13:45:53 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:25:46.418 13:45:53 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 57973' 00:25:46.418 13:45:53 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 57973 00:25:46.418 13:45:53 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 57973 00:25:48.953 13:45:55 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@90 -- # killprocess 57989 00:25:48.953 13:45:55 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 57989 ']' 00:25:48.953 13:45:55 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 57989 00:25:48.953 13:45:55 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:25:48.953 13:45:55 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:25:48.953 13:45:55 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 57989 00:25:48.953 killing process with pid 57989 00:25:48.953 13:45:55 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:25:48.953 13:45:55 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:25:48.953 13:45:55 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 57989' 00:25:48.953 13:45:55 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 57989 00:25:48.953 13:45:55 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 57989 00:25:50.326 00:25:50.326 real 0m6.908s 00:25:50.326 user 0m7.054s 00:25:50.326 sys 0m0.941s 00:25:50.326 13:45:57 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:25:50.326 ************************************ 00:25:50.326 END TEST non_locking_app_on_locked_coremask 00:25:50.326 13:45:57 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:25:50.326 ************************************ 00:25:50.326 13:45:57 event.cpu_locks -- event/cpu_locks.sh@169 -- # run_test locking_app_on_unlocked_coremask locking_app_on_unlocked_coremask 00:25:50.326 13:45:57 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:25:50.326 13:45:57 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:25:50.326 13:45:57 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:25:50.326 ************************************ 00:25:50.326 START TEST locking_app_on_unlocked_coremask 00:25:50.326 ************************************ 00:25:50.326 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:50.326 13:45:57 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1129 -- # locking_app_on_unlocked_coremask 00:25:50.326 13:45:57 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@98 -- # spdk_tgt_pid=58091 00:25:50.326 13:45:57 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@99 -- # waitforlisten 58091 /var/tmp/spdk.sock 00:25:50.326 13:45:57 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # '[' -z 58091 ']' 00:25:50.326 13:45:57 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:50.326 13:45:57 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:25:50.326 13:45:57 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:50.326 13:45:57 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:25:50.326 13:45:57 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:25:50.326 13:45:57 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@97 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks 00:25:50.326 [2024-11-20 13:45:57.280371] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:25:50.326 [2024-11-20 13:45:57.280488] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58091 ] 00:25:50.583 [2024-11-20 13:45:57.430055] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:25:50.583 [2024-11-20 13:45:57.430110] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:50.583 [2024-11-20 13:45:57.531846] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:51.150 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:25:51.150 13:45:58 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:25:51.150 13:45:58 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@868 -- # return 0 00:25:51.150 13:45:58 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@102 -- # spdk_tgt_pid2=58107 00:25:51.150 13:45:58 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@103 -- # waitforlisten 58107 /var/tmp/spdk2.sock 00:25:51.150 13:45:58 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # '[' -z 58107 ']' 00:25:51.150 13:45:58 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:25:51.150 13:45:58 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:25:51.150 13:45:58 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:25:51.150 13:45:58 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:25:51.150 13:45:58 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:25:51.150 13:45:58 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@101 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:25:51.408 [2024-11-20 13:45:58.273816] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:25:51.408 [2024-11-20 13:45:58.273960] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58107 ] 00:25:51.408 [2024-11-20 13:45:58.440248] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:51.665 [2024-11-20 13:45:58.641857] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:52.692 13:45:59 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:25:52.692 13:45:59 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@868 -- # return 0 00:25:52.692 13:45:59 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@105 -- # locks_exist 58107 00:25:52.692 13:45:59 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 58107 00:25:52.692 13:45:59 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:25:53.257 13:46:00 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@107 -- # killprocess 58091 00:25:53.257 13:46:00 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # '[' -z 58091 ']' 00:25:53.257 13:46:00 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@958 -- # kill -0 58091 00:25:53.257 13:46:00 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # uname 00:25:53.257 13:46:00 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:25:53.257 13:46:00 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 58091 00:25:53.257 killing process with pid 58091 00:25:53.257 13:46:00 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:25:53.257 13:46:00 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:25:53.257 13:46:00 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 58091' 00:25:53.257 13:46:00 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@973 -- # kill 58091 00:25:53.257 13:46:00 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@978 -- # wait 58091 00:25:55.784 13:46:02 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@108 -- # killprocess 58107 00:25:55.784 13:46:02 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # '[' -z 58107 ']' 00:25:55.784 13:46:02 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@958 -- # kill -0 58107 00:25:55.784 13:46:02 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # uname 00:25:55.784 13:46:02 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:25:55.784 13:46:02 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 58107 00:25:55.784 killing process with pid 58107 00:25:55.784 13:46:02 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:25:55.784 13:46:02 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:25:55.784 13:46:02 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 58107' 00:25:55.784 13:46:02 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@973 -- # kill 58107 00:25:55.784 13:46:02 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@978 -- # wait 58107 00:25:57.214 ************************************ 00:25:57.214 END TEST locking_app_on_unlocked_coremask 00:25:57.214 ************************************ 00:25:57.214 00:25:57.214 real 0m6.761s 00:25:57.214 user 0m6.986s 00:25:57.214 sys 0m0.955s 00:25:57.214 13:46:03 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:25:57.214 13:46:03 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:25:57.214 13:46:04 event.cpu_locks -- event/cpu_locks.sh@170 -- # run_test locking_app_on_locked_coremask locking_app_on_locked_coremask 00:25:57.214 13:46:04 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:25:57.214 13:46:04 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:25:57.214 13:46:04 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:25:57.214 ************************************ 00:25:57.214 START TEST locking_app_on_locked_coremask 00:25:57.214 ************************************ 00:25:57.214 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:57.214 13:46:04 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1129 -- # locking_app_on_locked_coremask 00:25:57.214 13:46:04 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@115 -- # spdk_tgt_pid=58204 00:25:57.214 13:46:04 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@116 -- # waitforlisten 58204 /var/tmp/spdk.sock 00:25:57.214 13:46:04 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@114 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:25:57.214 13:46:04 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 58204 ']' 00:25:57.214 13:46:04 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:57.214 13:46:04 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:25:57.214 13:46:04 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:57.214 13:46:04 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:25:57.214 13:46:04 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:25:57.214 [2024-11-20 13:46:04.085452] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:25:57.215 [2024-11-20 13:46:04.085571] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58204 ] 00:25:57.215 [2024-11-20 13:46:04.241211] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:57.474 [2024-11-20 13:46:04.359988] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:58.064 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:25:58.064 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:25:58.064 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@119 -- # spdk_tgt_pid2=58219 00:25:58.064 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@118 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:25:58.064 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@120 -- # NOT waitforlisten 58219 /var/tmp/spdk2.sock 00:25:58.064 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@652 -- # local es=0 00:25:58.064 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 58219 /var/tmp/spdk2.sock 00:25:58.064 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:25:58.064 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:25:58.064 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:25:58.064 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:25:58.064 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:25:58.064 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@655 -- # waitforlisten 58219 /var/tmp/spdk2.sock 00:25:58.064 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 58219 ']' 00:25:58.064 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:25:58.064 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:25:58.064 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:25:58.064 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:25:58.064 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:25:58.064 [2024-11-20 13:46:05.092942] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:25:58.064 [2024-11-20 13:46:05.093293] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58219 ] 00:25:58.322 [2024-11-20 13:46:05.268896] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 0, probably process 58204 has claimed it. 00:25:58.322 [2024-11-20 13:46:05.268971] app.c: 912:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:25:58.894 ERROR: process (pid: 58219) is no longer running 00:25:58.894 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 850: kill: (58219) - No such process 00:25:58.894 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:25:58.894 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 1 00:25:58.894 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@655 -- # es=1 00:25:58.894 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:25:58.894 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:25:58.894 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:25:58.894 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@122 -- # locks_exist 58204 00:25:58.894 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:25:58.894 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 58204 00:25:59.155 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@124 -- # killprocess 58204 00:25:59.155 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 58204 ']' 00:25:59.155 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 58204 00:25:59.155 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:25:59.155 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:25:59.155 13:46:05 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 58204 00:25:59.155 killing process with pid 58204 00:25:59.155 13:46:06 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:25:59.155 13:46:06 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:25:59.155 13:46:06 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 58204' 00:25:59.155 13:46:06 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 58204 00:25:59.155 13:46:06 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 58204 00:26:01.062 00:26:01.062 real 0m3.657s 00:26:01.062 user 0m3.844s 00:26:01.062 sys 0m0.642s 00:26:01.062 ************************************ 00:26:01.062 END TEST locking_app_on_locked_coremask 00:26:01.062 ************************************ 00:26:01.062 13:46:07 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:26:01.062 13:46:07 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:26:01.062 13:46:07 event.cpu_locks -- event/cpu_locks.sh@171 -- # run_test locking_overlapped_coremask locking_overlapped_coremask 00:26:01.062 13:46:07 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:26:01.062 13:46:07 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:26:01.062 13:46:07 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:26:01.062 ************************************ 00:26:01.062 START TEST locking_overlapped_coremask 00:26:01.062 ************************************ 00:26:01.062 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:01.062 13:46:07 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1129 -- # locking_overlapped_coremask 00:26:01.062 13:46:07 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@132 -- # spdk_tgt_pid=58278 00:26:01.062 13:46:07 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@131 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x7 00:26:01.062 13:46:07 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@133 -- # waitforlisten 58278 /var/tmp/spdk.sock 00:26:01.062 13:46:07 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # '[' -z 58278 ']' 00:26:01.062 13:46:07 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:01.062 13:46:07 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:01.062 13:46:07 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:01.062 13:46:07 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:01.062 13:46:07 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:26:01.062 [2024-11-20 13:46:07.798102] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:26:01.062 [2024-11-20 13:46:07.798744] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58278 ] 00:26:01.062 [2024-11-20 13:46:07.966270] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:26:01.062 [2024-11-20 13:46:08.094308] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:26:01.062 [2024-11-20 13:46:08.094389] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:26:01.062 [2024-11-20 13:46:08.094822] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:01.630 13:46:08 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:01.630 13:46:08 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@868 -- # return 0 00:26:01.630 13:46:08 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@136 -- # spdk_tgt_pid2=58296 00:26:01.630 13:46:08 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@137 -- # NOT waitforlisten 58296 /var/tmp/spdk2.sock 00:26:01.630 13:46:08 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@652 -- # local es=0 00:26:01.630 13:46:08 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@135 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock 00:26:01.630 13:46:08 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 58296 /var/tmp/spdk2.sock 00:26:01.630 13:46:08 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:26:01.630 13:46:08 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:01.630 13:46:08 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:26:01.630 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:26:01.630 13:46:08 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:01.630 13:46:08 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@655 -- # waitforlisten 58296 /var/tmp/spdk2.sock 00:26:01.630 13:46:08 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # '[' -z 58296 ']' 00:26:01.630 13:46:08 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:26:01.630 13:46:08 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:01.630 13:46:08 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:26:01.630 13:46:08 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:01.630 13:46:08 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:26:01.887 [2024-11-20 13:46:08.715761] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:26:01.887 [2024-11-20 13:46:08.715893] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58296 ] 00:26:01.887 [2024-11-20 13:46:08.894499] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 58278 has claimed it. 00:26:01.887 [2024-11-20 13:46:08.894578] app.c: 912:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:26:02.451 ERROR: process (pid: 58296) is no longer running 00:26:02.451 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 850: kill: (58296) - No such process 00:26:02.451 13:46:09 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:02.451 13:46:09 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@868 -- # return 1 00:26:02.451 13:46:09 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@655 -- # es=1 00:26:02.451 13:46:09 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:26:02.451 13:46:09 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:26:02.451 13:46:09 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:26:02.451 13:46:09 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@139 -- # check_remaining_locks 00:26:02.451 13:46:09 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:26:02.451 13:46:09 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:26:02.451 13:46:09 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:26:02.451 13:46:09 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@141 -- # killprocess 58278 00:26:02.451 13:46:09 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@954 -- # '[' -z 58278 ']' 00:26:02.451 13:46:09 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@958 -- # kill -0 58278 00:26:02.451 13:46:09 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@959 -- # uname 00:26:02.451 13:46:09 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:02.451 13:46:09 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 58278 00:26:02.451 13:46:09 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:26:02.451 killing process with pid 58278 00:26:02.451 13:46:09 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:26:02.451 13:46:09 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 58278' 00:26:02.451 13:46:09 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@973 -- # kill 58278 00:26:02.451 13:46:09 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@978 -- # wait 58278 00:26:03.828 00:26:03.828 real 0m2.985s 00:26:03.828 user 0m7.935s 00:26:03.828 sys 0m0.494s 00:26:03.828 ************************************ 00:26:03.828 END TEST locking_overlapped_coremask 00:26:03.828 ************************************ 00:26:03.828 13:46:10 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:26:03.828 13:46:10 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:26:03.828 13:46:10 event.cpu_locks -- event/cpu_locks.sh@172 -- # run_test locking_overlapped_coremask_via_rpc locking_overlapped_coremask_via_rpc 00:26:03.828 13:46:10 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:26:03.828 13:46:10 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:26:03.828 13:46:10 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:26:03.828 ************************************ 00:26:03.828 START TEST locking_overlapped_coremask_via_rpc 00:26:03.828 ************************************ 00:26:03.828 13:46:10 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1129 -- # locking_overlapped_coremask_via_rpc 00:26:03.828 13:46:10 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@148 -- # spdk_tgt_pid=58349 00:26:03.828 13:46:10 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@149 -- # waitforlisten 58349 /var/tmp/spdk.sock 00:26:03.828 13:46:10 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 58349 ']' 00:26:03.828 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:03.828 13:46:10 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:03.828 13:46:10 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:03.828 13:46:10 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:03.828 13:46:10 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:03.828 13:46:10 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@147 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x7 --disable-cpumask-locks 00:26:03.828 13:46:10 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:26:03.828 [2024-11-20 13:46:10.823500] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:26:03.828 [2024-11-20 13:46:10.823629] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58349 ] 00:26:04.087 [2024-11-20 13:46:10.986776] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:26:04.087 [2024-11-20 13:46:10.986857] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:26:04.087 [2024-11-20 13:46:11.110355] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:26:04.087 [2024-11-20 13:46:11.111293] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:26:04.087 [2024-11-20 13:46:11.111451] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:05.076 13:46:11 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:05.076 13:46:11 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:26:05.076 13:46:11 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@152 -- # spdk_tgt_pid2=58367 00:26:05.076 13:46:11 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@153 -- # waitforlisten 58367 /var/tmp/spdk2.sock 00:26:05.077 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:26:05.077 13:46:11 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 58367 ']' 00:26:05.077 13:46:11 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:26:05.077 13:46:11 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:05.077 13:46:11 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:26:05.077 13:46:11 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:05.077 13:46:11 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:26:05.077 13:46:11 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@151 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock --disable-cpumask-locks 00:26:05.077 [2024-11-20 13:46:11.906789] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:26:05.077 [2024-11-20 13:46:11.906921] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58367 ] 00:26:05.077 [2024-11-20 13:46:12.073520] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:26:05.077 [2024-11-20 13:46:12.073588] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:26:05.334 [2024-11-20 13:46:12.288773] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:26:05.334 [2024-11-20 13:46:12.288845] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:26:05.334 [2024-11-20 13:46:12.288862] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@155 -- # rpc_cmd framework_enable_cpumask_locks 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@156 -- # NOT rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@652 -- # local es=0 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@655 -- # rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:26:06.708 [2024-11-20 13:46:13.384156] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 58349 has claimed it. 00:26:06.708 request: 00:26:06.708 { 00:26:06.708 "method": "framework_enable_cpumask_locks", 00:26:06.708 "req_id": 1 00:26:06.708 } 00:26:06.708 Got JSON-RPC error response 00:26:06.708 response: 00:26:06.708 { 00:26:06.708 "code": -32603, 00:26:06.708 "message": "Failed to claim CPU core: 2" 00:26:06.708 } 00:26:06.708 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@655 -- # es=1 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@158 -- # waitforlisten 58349 /var/tmp/spdk.sock 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 58349 ']' 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:06.708 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:26:06.708 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:26:06.709 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:06.709 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:26:06.709 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@159 -- # waitforlisten 58367 /var/tmp/spdk2.sock 00:26:06.709 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 58367 ']' 00:26:06.709 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:26:06.709 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:06.709 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:26:06.709 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:06.709 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:26:06.969 ************************************ 00:26:06.969 END TEST locking_overlapped_coremask_via_rpc 00:26:06.969 ************************************ 00:26:06.969 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:06.969 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:26:06.969 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@161 -- # check_remaining_locks 00:26:06.969 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:26:06.969 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:26:06.969 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:26:06.969 00:26:06.969 real 0m3.124s 00:26:06.969 user 0m1.271s 00:26:06.969 sys 0m0.134s 00:26:06.969 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:26:06.969 13:46:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:26:06.969 13:46:13 event.cpu_locks -- event/cpu_locks.sh@174 -- # cleanup 00:26:06.969 13:46:13 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 58349 ]] 00:26:06.969 13:46:13 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 58349 00:26:06.969 13:46:13 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 58349 ']' 00:26:06.969 13:46:13 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 58349 00:26:06.969 13:46:13 event.cpu_locks -- common/autotest_common.sh@959 -- # uname 00:26:06.969 13:46:13 event.cpu_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:06.969 13:46:13 event.cpu_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 58349 00:26:06.969 killing process with pid 58349 00:26:06.969 13:46:13 event.cpu_locks -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:26:06.969 13:46:13 event.cpu_locks -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:26:06.969 13:46:13 event.cpu_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 58349' 00:26:06.969 13:46:13 event.cpu_locks -- common/autotest_common.sh@973 -- # kill 58349 00:26:06.969 13:46:13 event.cpu_locks -- common/autotest_common.sh@978 -- # wait 58349 00:26:08.347 13:46:15 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 58367 ]] 00:26:08.347 13:46:15 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 58367 00:26:08.347 13:46:15 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 58367 ']' 00:26:08.347 13:46:15 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 58367 00:26:08.347 13:46:15 event.cpu_locks -- common/autotest_common.sh@959 -- # uname 00:26:08.347 13:46:15 event.cpu_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:08.347 13:46:15 event.cpu_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 58367 00:26:08.347 killing process with pid 58367 00:26:08.347 13:46:15 event.cpu_locks -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:26:08.347 13:46:15 event.cpu_locks -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:26:08.347 13:46:15 event.cpu_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 58367' 00:26:08.347 13:46:15 event.cpu_locks -- common/autotest_common.sh@973 -- # kill 58367 00:26:08.347 13:46:15 event.cpu_locks -- common/autotest_common.sh@978 -- # wait 58367 00:26:09.721 13:46:16 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:26:09.721 Process with pid 58349 is not found 00:26:09.721 13:46:16 event.cpu_locks -- event/cpu_locks.sh@1 -- # cleanup 00:26:09.721 13:46:16 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 58349 ]] 00:26:09.721 13:46:16 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 58349 00:26:09.721 13:46:16 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 58349 ']' 00:26:09.721 13:46:16 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 58349 00:26:09.722 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 958: kill: (58349) - No such process 00:26:09.722 13:46:16 event.cpu_locks -- common/autotest_common.sh@981 -- # echo 'Process with pid 58349 is not found' 00:26:09.722 13:46:16 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 58367 ]] 00:26:09.722 13:46:16 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 58367 00:26:09.722 Process with pid 58367 is not found 00:26:09.722 13:46:16 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 58367 ']' 00:26:09.722 13:46:16 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 58367 00:26:09.722 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 958: kill: (58367) - No such process 00:26:09.722 13:46:16 event.cpu_locks -- common/autotest_common.sh@981 -- # echo 'Process with pid 58367 is not found' 00:26:09.722 13:46:16 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:26:09.980 ************************************ 00:26:09.980 END TEST cpu_locks 00:26:09.980 ************************************ 00:26:09.980 00:26:09.980 real 0m32.048s 00:26:09.980 user 0m54.682s 00:26:09.980 sys 0m5.076s 00:26:09.980 13:46:16 event.cpu_locks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:26:09.980 13:46:16 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:26:09.980 ************************************ 00:26:09.980 END TEST event 00:26:09.980 ************************************ 00:26:09.980 00:26:09.980 real 0m58.124s 00:26:09.980 user 1m46.949s 00:26:09.980 sys 0m8.168s 00:26:09.980 13:46:16 event -- common/autotest_common.sh@1130 -- # xtrace_disable 00:26:09.980 13:46:16 event -- common/autotest_common.sh@10 -- # set +x 00:26:09.980 13:46:16 -- spdk/autotest.sh@169 -- # run_test thread /home/vagrant/spdk_repo/spdk/test/thread/thread.sh 00:26:09.980 13:46:16 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:26:09.980 13:46:16 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:26:09.980 13:46:16 -- common/autotest_common.sh@10 -- # set +x 00:26:09.980 ************************************ 00:26:09.980 START TEST thread 00:26:09.980 ************************************ 00:26:09.980 13:46:16 thread -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/thread/thread.sh 00:26:09.980 * Looking for test storage... 00:26:09.980 * Found test storage at /home/vagrant/spdk_repo/spdk/test/thread 00:26:09.980 13:46:16 thread -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:26:09.980 13:46:16 thread -- common/autotest_common.sh@1693 -- # lcov --version 00:26:09.980 13:46:16 thread -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:26:09.980 13:46:16 thread -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:26:09.980 13:46:16 thread -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:26:09.980 13:46:16 thread -- scripts/common.sh@333 -- # local ver1 ver1_l 00:26:09.980 13:46:16 thread -- scripts/common.sh@334 -- # local ver2 ver2_l 00:26:09.980 13:46:16 thread -- scripts/common.sh@336 -- # IFS=.-: 00:26:09.980 13:46:16 thread -- scripts/common.sh@336 -- # read -ra ver1 00:26:09.980 13:46:16 thread -- scripts/common.sh@337 -- # IFS=.-: 00:26:09.980 13:46:16 thread -- scripts/common.sh@337 -- # read -ra ver2 00:26:09.980 13:46:16 thread -- scripts/common.sh@338 -- # local 'op=<' 00:26:09.980 13:46:16 thread -- scripts/common.sh@340 -- # ver1_l=2 00:26:09.980 13:46:16 thread -- scripts/common.sh@341 -- # ver2_l=1 00:26:09.980 13:46:16 thread -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:26:09.980 13:46:16 thread -- scripts/common.sh@344 -- # case "$op" in 00:26:09.980 13:46:16 thread -- scripts/common.sh@345 -- # : 1 00:26:09.980 13:46:16 thread -- scripts/common.sh@364 -- # (( v = 0 )) 00:26:09.980 13:46:16 thread -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:26:09.980 13:46:16 thread -- scripts/common.sh@365 -- # decimal 1 00:26:09.980 13:46:16 thread -- scripts/common.sh@353 -- # local d=1 00:26:09.980 13:46:16 thread -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:26:09.980 13:46:16 thread -- scripts/common.sh@355 -- # echo 1 00:26:09.980 13:46:16 thread -- scripts/common.sh@365 -- # ver1[v]=1 00:26:09.980 13:46:16 thread -- scripts/common.sh@366 -- # decimal 2 00:26:09.980 13:46:16 thread -- scripts/common.sh@353 -- # local d=2 00:26:09.980 13:46:16 thread -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:26:09.980 13:46:16 thread -- scripts/common.sh@355 -- # echo 2 00:26:09.980 13:46:16 thread -- scripts/common.sh@366 -- # ver2[v]=2 00:26:09.980 13:46:16 thread -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:26:09.980 13:46:16 thread -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:26:09.980 13:46:16 thread -- scripts/common.sh@368 -- # return 0 00:26:09.980 13:46:16 thread -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:26:09.980 13:46:16 thread -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:26:09.980 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:09.980 --rc genhtml_branch_coverage=1 00:26:09.980 --rc genhtml_function_coverage=1 00:26:09.980 --rc genhtml_legend=1 00:26:09.980 --rc geninfo_all_blocks=1 00:26:09.980 --rc geninfo_unexecuted_blocks=1 00:26:09.980 00:26:09.980 ' 00:26:09.980 13:46:16 thread -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:26:09.980 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:09.980 --rc genhtml_branch_coverage=1 00:26:09.980 --rc genhtml_function_coverage=1 00:26:09.980 --rc genhtml_legend=1 00:26:09.980 --rc geninfo_all_blocks=1 00:26:09.980 --rc geninfo_unexecuted_blocks=1 00:26:09.980 00:26:09.980 ' 00:26:09.980 13:46:16 thread -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:26:09.980 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:09.980 --rc genhtml_branch_coverage=1 00:26:09.980 --rc genhtml_function_coverage=1 00:26:09.980 --rc genhtml_legend=1 00:26:09.980 --rc geninfo_all_blocks=1 00:26:09.980 --rc geninfo_unexecuted_blocks=1 00:26:09.980 00:26:09.980 ' 00:26:09.981 13:46:16 thread -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:26:09.981 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:09.981 --rc genhtml_branch_coverage=1 00:26:09.981 --rc genhtml_function_coverage=1 00:26:09.981 --rc genhtml_legend=1 00:26:09.981 --rc geninfo_all_blocks=1 00:26:09.981 --rc geninfo_unexecuted_blocks=1 00:26:09.981 00:26:09.981 ' 00:26:09.981 13:46:16 thread -- thread/thread.sh@11 -- # run_test thread_poller_perf /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:26:09.981 13:46:16 thread -- common/autotest_common.sh@1105 -- # '[' 8 -le 1 ']' 00:26:09.981 13:46:16 thread -- common/autotest_common.sh@1111 -- # xtrace_disable 00:26:09.981 13:46:16 thread -- common/autotest_common.sh@10 -- # set +x 00:26:09.981 ************************************ 00:26:09.981 START TEST thread_poller_perf 00:26:09.981 ************************************ 00:26:09.981 13:46:17 thread.thread_poller_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:26:09.981 [2024-11-20 13:46:17.033544] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:26:09.981 [2024-11-20 13:46:17.033822] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58527 ] 00:26:10.238 [2024-11-20 13:46:17.195152] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:10.496 [2024-11-20 13:46:17.316508] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:10.496 Running 1000 pollers for 1 seconds with 1 microseconds period. 00:26:11.429 [2024-11-20T13:46:18.488Z] ====================================== 00:26:11.429 [2024-11-20T13:46:18.488Z] busy:2615839684 (cyc) 00:26:11.429 [2024-11-20T13:46:18.488Z] total_run_count: 305000 00:26:11.429 [2024-11-20T13:46:18.488Z] tsc_hz: 2600000000 (cyc) 00:26:11.429 [2024-11-20T13:46:18.488Z] ====================================== 00:26:11.429 [2024-11-20T13:46:18.488Z] poller_cost: 8576 (cyc), 3298 (nsec) 00:26:11.687 00:26:11.687 real 0m1.488s 00:26:11.687 user 0m1.307s 00:26:11.687 sys 0m0.072s 00:26:11.687 13:46:18 thread.thread_poller_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:26:11.687 13:46:18 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:26:11.687 ************************************ 00:26:11.687 END TEST thread_poller_perf 00:26:11.687 ************************************ 00:26:11.687 13:46:18 thread -- thread/thread.sh@12 -- # run_test thread_poller_perf /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:26:11.687 13:46:18 thread -- common/autotest_common.sh@1105 -- # '[' 8 -le 1 ']' 00:26:11.687 13:46:18 thread -- common/autotest_common.sh@1111 -- # xtrace_disable 00:26:11.687 13:46:18 thread -- common/autotest_common.sh@10 -- # set +x 00:26:11.687 ************************************ 00:26:11.687 START TEST thread_poller_perf 00:26:11.687 ************************************ 00:26:11.687 13:46:18 thread.thread_poller_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:26:11.687 [2024-11-20 13:46:18.576022] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:26:11.687 [2024-11-20 13:46:18.576351] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58564 ] 00:26:11.687 [2024-11-20 13:46:18.734872] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:11.945 [2024-11-20 13:46:18.854618] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:11.946 Running 1000 pollers for 1 seconds with 0 microseconds period. 00:26:13.320 [2024-11-20T13:46:20.379Z] ====================================== 00:26:13.320 [2024-11-20T13:46:20.379Z] busy:2603658104 (cyc) 00:26:13.320 [2024-11-20T13:46:20.379Z] total_run_count: 3968000 00:26:13.320 [2024-11-20T13:46:20.379Z] tsc_hz: 2600000000 (cyc) 00:26:13.320 [2024-11-20T13:46:20.379Z] ====================================== 00:26:13.320 [2024-11-20T13:46:20.379Z] poller_cost: 656 (cyc), 252 (nsec) 00:26:13.320 00:26:13.320 real 0m1.478s 00:26:13.320 user 0m1.295s 00:26:13.320 sys 0m0.074s 00:26:13.320 13:46:20 thread.thread_poller_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:26:13.320 ************************************ 00:26:13.320 END TEST thread_poller_perf 00:26:13.320 ************************************ 00:26:13.320 13:46:20 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:26:13.320 13:46:20 thread -- thread/thread.sh@17 -- # [[ y != \y ]] 00:26:13.320 00:26:13.320 real 0m3.206s 00:26:13.320 user 0m2.724s 00:26:13.320 sys 0m0.263s 00:26:13.320 ************************************ 00:26:13.320 END TEST thread 00:26:13.320 ************************************ 00:26:13.320 13:46:20 thread -- common/autotest_common.sh@1130 -- # xtrace_disable 00:26:13.320 13:46:20 thread -- common/autotest_common.sh@10 -- # set +x 00:26:13.320 13:46:20 -- spdk/autotest.sh@171 -- # [[ 0 -eq 1 ]] 00:26:13.320 13:46:20 -- spdk/autotest.sh@176 -- # run_test app_cmdline /home/vagrant/spdk_repo/spdk/test/app/cmdline.sh 00:26:13.320 13:46:20 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:26:13.320 13:46:20 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:26:13.320 13:46:20 -- common/autotest_common.sh@10 -- # set +x 00:26:13.320 ************************************ 00:26:13.320 START TEST app_cmdline 00:26:13.320 ************************************ 00:26:13.320 13:46:20 app_cmdline -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/app/cmdline.sh 00:26:13.320 * Looking for test storage... 00:26:13.320 * Found test storage at /home/vagrant/spdk_repo/spdk/test/app 00:26:13.320 13:46:20 app_cmdline -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:26:13.320 13:46:20 app_cmdline -- common/autotest_common.sh@1693 -- # lcov --version 00:26:13.321 13:46:20 app_cmdline -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:26:13.321 13:46:20 app_cmdline -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@333 -- # local ver1 ver1_l 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@334 -- # local ver2 ver2_l 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@336 -- # IFS=.-: 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@336 -- # read -ra ver1 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@337 -- # IFS=.-: 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@337 -- # read -ra ver2 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@338 -- # local 'op=<' 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@340 -- # ver1_l=2 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@341 -- # ver2_l=1 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@344 -- # case "$op" in 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@345 -- # : 1 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@364 -- # (( v = 0 )) 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@365 -- # decimal 1 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@353 -- # local d=1 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@355 -- # echo 1 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@365 -- # ver1[v]=1 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@366 -- # decimal 2 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@353 -- # local d=2 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@355 -- # echo 2 00:26:13.321 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@366 -- # ver2[v]=2 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:26:13.321 13:46:20 app_cmdline -- scripts/common.sh@368 -- # return 0 00:26:13.321 13:46:20 app_cmdline -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:26:13.321 13:46:20 app_cmdline -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:26:13.321 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:13.321 --rc genhtml_branch_coverage=1 00:26:13.321 --rc genhtml_function_coverage=1 00:26:13.321 --rc genhtml_legend=1 00:26:13.321 --rc geninfo_all_blocks=1 00:26:13.321 --rc geninfo_unexecuted_blocks=1 00:26:13.321 00:26:13.321 ' 00:26:13.321 13:46:20 app_cmdline -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:26:13.321 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:13.321 --rc genhtml_branch_coverage=1 00:26:13.321 --rc genhtml_function_coverage=1 00:26:13.321 --rc genhtml_legend=1 00:26:13.321 --rc geninfo_all_blocks=1 00:26:13.321 --rc geninfo_unexecuted_blocks=1 00:26:13.321 00:26:13.321 ' 00:26:13.321 13:46:20 app_cmdline -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:26:13.321 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:13.321 --rc genhtml_branch_coverage=1 00:26:13.321 --rc genhtml_function_coverage=1 00:26:13.321 --rc genhtml_legend=1 00:26:13.321 --rc geninfo_all_blocks=1 00:26:13.321 --rc geninfo_unexecuted_blocks=1 00:26:13.321 00:26:13.321 ' 00:26:13.321 13:46:20 app_cmdline -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:26:13.321 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:13.321 --rc genhtml_branch_coverage=1 00:26:13.321 --rc genhtml_function_coverage=1 00:26:13.321 --rc genhtml_legend=1 00:26:13.321 --rc geninfo_all_blocks=1 00:26:13.321 --rc geninfo_unexecuted_blocks=1 00:26:13.321 00:26:13.321 ' 00:26:13.321 13:46:20 app_cmdline -- app/cmdline.sh@14 -- # trap 'killprocess $spdk_tgt_pid' EXIT 00:26:13.321 13:46:20 app_cmdline -- app/cmdline.sh@17 -- # spdk_tgt_pid=58647 00:26:13.321 13:46:20 app_cmdline -- app/cmdline.sh@18 -- # waitforlisten 58647 00:26:13.321 13:46:20 app_cmdline -- common/autotest_common.sh@835 -- # '[' -z 58647 ']' 00:26:13.321 13:46:20 app_cmdline -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:13.321 13:46:20 app_cmdline -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:13.321 13:46:20 app_cmdline -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:13.321 13:46:20 app_cmdline -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:13.321 13:46:20 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:26:13.321 13:46:20 app_cmdline -- app/cmdline.sh@16 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --rpcs-allowed spdk_get_version,rpc_get_methods 00:26:13.321 [2024-11-20 13:46:20.308880] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:26:13.321 [2024-11-20 13:46:20.309017] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58647 ] 00:26:13.578 [2024-11-20 13:46:20.466595] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:13.578 [2024-11-20 13:46:20.586848] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:14.513 13:46:21 app_cmdline -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:14.513 13:46:21 app_cmdline -- common/autotest_common.sh@868 -- # return 0 00:26:14.513 13:46:21 app_cmdline -- app/cmdline.sh@20 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py spdk_get_version 00:26:14.513 { 00:26:14.513 "version": "SPDK v25.01-pre git sha1 a361eb5e2", 00:26:14.513 "fields": { 00:26:14.513 "major": 25, 00:26:14.513 "minor": 1, 00:26:14.513 "patch": 0, 00:26:14.513 "suffix": "-pre", 00:26:14.513 "commit": "a361eb5e2" 00:26:14.513 } 00:26:14.513 } 00:26:14.513 13:46:21 app_cmdline -- app/cmdline.sh@22 -- # expected_methods=() 00:26:14.513 13:46:21 app_cmdline -- app/cmdline.sh@23 -- # expected_methods+=("rpc_get_methods") 00:26:14.513 13:46:21 app_cmdline -- app/cmdline.sh@24 -- # expected_methods+=("spdk_get_version") 00:26:14.513 13:46:21 app_cmdline -- app/cmdline.sh@26 -- # methods=($(rpc_cmd rpc_get_methods | jq -r ".[]" | sort)) 00:26:14.513 13:46:21 app_cmdline -- app/cmdline.sh@26 -- # rpc_cmd rpc_get_methods 00:26:14.513 13:46:21 app_cmdline -- app/cmdline.sh@26 -- # jq -r '.[]' 00:26:14.513 13:46:21 app_cmdline -- app/cmdline.sh@26 -- # sort 00:26:14.513 13:46:21 app_cmdline -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:14.513 13:46:21 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:26:14.513 13:46:21 app_cmdline -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:14.513 13:46:21 app_cmdline -- app/cmdline.sh@27 -- # (( 2 == 2 )) 00:26:14.513 13:46:21 app_cmdline -- app/cmdline.sh@28 -- # [[ rpc_get_methods spdk_get_version == \r\p\c\_\g\e\t\_\m\e\t\h\o\d\s\ \s\p\d\k\_\g\e\t\_\v\e\r\s\i\o\n ]] 00:26:14.513 13:46:21 app_cmdline -- app/cmdline.sh@30 -- # NOT /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:26:14.513 13:46:21 app_cmdline -- common/autotest_common.sh@652 -- # local es=0 00:26:14.513 13:46:21 app_cmdline -- common/autotest_common.sh@654 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:26:14.513 13:46:21 app_cmdline -- common/autotest_common.sh@640 -- # local arg=/home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:26:14.513 13:46:21 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:14.513 13:46:21 app_cmdline -- common/autotest_common.sh@644 -- # type -t /home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:26:14.513 13:46:21 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:14.513 13:46:21 app_cmdline -- common/autotest_common.sh@646 -- # type -P /home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:26:14.513 13:46:21 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:14.513 13:46:21 app_cmdline -- common/autotest_common.sh@646 -- # arg=/home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:26:14.513 13:46:21 app_cmdline -- common/autotest_common.sh@646 -- # [[ -x /home/vagrant/spdk_repo/spdk/scripts/rpc.py ]] 00:26:14.513 13:46:21 app_cmdline -- common/autotest_common.sh@655 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:26:14.771 request: 00:26:14.771 { 00:26:14.771 "method": "env_dpdk_get_mem_stats", 00:26:14.771 "req_id": 1 00:26:14.771 } 00:26:14.771 Got JSON-RPC error response 00:26:14.771 response: 00:26:14.771 { 00:26:14.771 "code": -32601, 00:26:14.771 "message": "Method not found" 00:26:14.771 } 00:26:14.771 13:46:21 app_cmdline -- common/autotest_common.sh@655 -- # es=1 00:26:14.771 13:46:21 app_cmdline -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:26:14.771 13:46:21 app_cmdline -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:26:14.771 13:46:21 app_cmdline -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:26:14.771 13:46:21 app_cmdline -- app/cmdline.sh@1 -- # killprocess 58647 00:26:14.771 13:46:21 app_cmdline -- common/autotest_common.sh@954 -- # '[' -z 58647 ']' 00:26:14.771 13:46:21 app_cmdline -- common/autotest_common.sh@958 -- # kill -0 58647 00:26:14.772 13:46:21 app_cmdline -- common/autotest_common.sh@959 -- # uname 00:26:14.772 13:46:21 app_cmdline -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:14.772 13:46:21 app_cmdline -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 58647 00:26:14.772 killing process with pid 58647 00:26:14.772 13:46:21 app_cmdline -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:26:14.772 13:46:21 app_cmdline -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:26:14.772 13:46:21 app_cmdline -- common/autotest_common.sh@972 -- # echo 'killing process with pid 58647' 00:26:14.772 13:46:21 app_cmdline -- common/autotest_common.sh@973 -- # kill 58647 00:26:14.772 13:46:21 app_cmdline -- common/autotest_common.sh@978 -- # wait 58647 00:26:16.723 ************************************ 00:26:16.723 END TEST app_cmdline 00:26:16.723 ************************************ 00:26:16.723 00:26:16.723 real 0m3.178s 00:26:16.723 user 0m3.342s 00:26:16.723 sys 0m0.524s 00:26:16.723 13:46:23 app_cmdline -- common/autotest_common.sh@1130 -- # xtrace_disable 00:26:16.723 13:46:23 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:26:16.723 13:46:23 -- spdk/autotest.sh@177 -- # run_test version /home/vagrant/spdk_repo/spdk/test/app/version.sh 00:26:16.723 13:46:23 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:26:16.723 13:46:23 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:26:16.723 13:46:23 -- common/autotest_common.sh@10 -- # set +x 00:26:16.723 ************************************ 00:26:16.723 START TEST version 00:26:16.723 ************************************ 00:26:16.723 13:46:23 version -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/app/version.sh 00:26:16.724 * Looking for test storage... 00:26:16.724 * Found test storage at /home/vagrant/spdk_repo/spdk/test/app 00:26:16.724 13:46:23 version -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:26:16.724 13:46:23 version -- common/autotest_common.sh@1693 -- # lcov --version 00:26:16.724 13:46:23 version -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:26:16.724 13:46:23 version -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:26:16.724 13:46:23 version -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:26:16.724 13:46:23 version -- scripts/common.sh@333 -- # local ver1 ver1_l 00:26:16.724 13:46:23 version -- scripts/common.sh@334 -- # local ver2 ver2_l 00:26:16.724 13:46:23 version -- scripts/common.sh@336 -- # IFS=.-: 00:26:16.724 13:46:23 version -- scripts/common.sh@336 -- # read -ra ver1 00:26:16.724 13:46:23 version -- scripts/common.sh@337 -- # IFS=.-: 00:26:16.724 13:46:23 version -- scripts/common.sh@337 -- # read -ra ver2 00:26:16.724 13:46:23 version -- scripts/common.sh@338 -- # local 'op=<' 00:26:16.724 13:46:23 version -- scripts/common.sh@340 -- # ver1_l=2 00:26:16.724 13:46:23 version -- scripts/common.sh@341 -- # ver2_l=1 00:26:16.724 13:46:23 version -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:26:16.724 13:46:23 version -- scripts/common.sh@344 -- # case "$op" in 00:26:16.724 13:46:23 version -- scripts/common.sh@345 -- # : 1 00:26:16.724 13:46:23 version -- scripts/common.sh@364 -- # (( v = 0 )) 00:26:16.724 13:46:23 version -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:26:16.724 13:46:23 version -- scripts/common.sh@365 -- # decimal 1 00:26:16.724 13:46:23 version -- scripts/common.sh@353 -- # local d=1 00:26:16.724 13:46:23 version -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:26:16.724 13:46:23 version -- scripts/common.sh@355 -- # echo 1 00:26:16.724 13:46:23 version -- scripts/common.sh@365 -- # ver1[v]=1 00:26:16.724 13:46:23 version -- scripts/common.sh@366 -- # decimal 2 00:26:16.724 13:46:23 version -- scripts/common.sh@353 -- # local d=2 00:26:16.724 13:46:23 version -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:26:16.724 13:46:23 version -- scripts/common.sh@355 -- # echo 2 00:26:16.724 13:46:23 version -- scripts/common.sh@366 -- # ver2[v]=2 00:26:16.724 13:46:23 version -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:26:16.724 13:46:23 version -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:26:16.724 13:46:23 version -- scripts/common.sh@368 -- # return 0 00:26:16.724 13:46:23 version -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:26:16.724 13:46:23 version -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:26:16.724 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:16.724 --rc genhtml_branch_coverage=1 00:26:16.724 --rc genhtml_function_coverage=1 00:26:16.724 --rc genhtml_legend=1 00:26:16.724 --rc geninfo_all_blocks=1 00:26:16.724 --rc geninfo_unexecuted_blocks=1 00:26:16.724 00:26:16.724 ' 00:26:16.724 13:46:23 version -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:26:16.724 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:16.724 --rc genhtml_branch_coverage=1 00:26:16.724 --rc genhtml_function_coverage=1 00:26:16.724 --rc genhtml_legend=1 00:26:16.724 --rc geninfo_all_blocks=1 00:26:16.724 --rc geninfo_unexecuted_blocks=1 00:26:16.724 00:26:16.724 ' 00:26:16.724 13:46:23 version -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:26:16.724 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:16.724 --rc genhtml_branch_coverage=1 00:26:16.724 --rc genhtml_function_coverage=1 00:26:16.724 --rc genhtml_legend=1 00:26:16.724 --rc geninfo_all_blocks=1 00:26:16.724 --rc geninfo_unexecuted_blocks=1 00:26:16.724 00:26:16.724 ' 00:26:16.724 13:46:23 version -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:26:16.724 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:16.724 --rc genhtml_branch_coverage=1 00:26:16.724 --rc genhtml_function_coverage=1 00:26:16.724 --rc genhtml_legend=1 00:26:16.724 --rc geninfo_all_blocks=1 00:26:16.724 --rc geninfo_unexecuted_blocks=1 00:26:16.724 00:26:16.724 ' 00:26:16.724 13:46:23 version -- app/version.sh@17 -- # get_header_version major 00:26:16.724 13:46:23 version -- app/version.sh@14 -- # cut -f2 00:26:16.724 13:46:23 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MAJOR[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:26:16.724 13:46:23 version -- app/version.sh@14 -- # tr -d '"' 00:26:16.724 13:46:23 version -- app/version.sh@17 -- # major=25 00:26:16.724 13:46:23 version -- app/version.sh@18 -- # get_header_version minor 00:26:16.724 13:46:23 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MINOR[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:26:16.724 13:46:23 version -- app/version.sh@14 -- # tr -d '"' 00:26:16.724 13:46:23 version -- app/version.sh@14 -- # cut -f2 00:26:16.724 13:46:23 version -- app/version.sh@18 -- # minor=1 00:26:16.724 13:46:23 version -- app/version.sh@19 -- # get_header_version patch 00:26:16.724 13:46:23 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_PATCH[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:26:16.724 13:46:23 version -- app/version.sh@14 -- # tr -d '"' 00:26:16.724 13:46:23 version -- app/version.sh@14 -- # cut -f2 00:26:16.724 13:46:23 version -- app/version.sh@19 -- # patch=0 00:26:16.724 13:46:23 version -- app/version.sh@20 -- # get_header_version suffix 00:26:16.724 13:46:23 version -- app/version.sh@14 -- # cut -f2 00:26:16.724 13:46:23 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_SUFFIX[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:26:16.724 13:46:23 version -- app/version.sh@14 -- # tr -d '"' 00:26:16.724 13:46:23 version -- app/version.sh@20 -- # suffix=-pre 00:26:16.724 13:46:23 version -- app/version.sh@22 -- # version=25.1 00:26:16.724 13:46:23 version -- app/version.sh@25 -- # (( patch != 0 )) 00:26:16.724 13:46:23 version -- app/version.sh@28 -- # version=25.1rc0 00:26:16.724 13:46:23 version -- app/version.sh@30 -- # PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python 00:26:16.724 13:46:23 version -- app/version.sh@30 -- # python3 -c 'import spdk; print(spdk.__version__)' 00:26:16.724 13:46:23 version -- app/version.sh@30 -- # py_version=25.1rc0 00:26:16.724 13:46:23 version -- app/version.sh@31 -- # [[ 25.1rc0 == \2\5\.\1\r\c\0 ]] 00:26:16.724 00:26:16.724 real 0m0.193s 00:26:16.724 user 0m0.122s 00:26:16.724 sys 0m0.097s 00:26:16.724 ************************************ 00:26:16.724 END TEST version 00:26:16.724 ************************************ 00:26:16.724 13:46:23 version -- common/autotest_common.sh@1130 -- # xtrace_disable 00:26:16.724 13:46:23 version -- common/autotest_common.sh@10 -- # set +x 00:26:16.724 13:46:23 -- spdk/autotest.sh@179 -- # '[' 0 -eq 1 ']' 00:26:16.724 13:46:23 -- spdk/autotest.sh@188 -- # [[ 1 -eq 1 ]] 00:26:16.724 13:46:23 -- spdk/autotest.sh@189 -- # run_test bdev_raid /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh 00:26:16.724 13:46:23 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:26:16.724 13:46:23 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:26:16.724 13:46:23 -- common/autotest_common.sh@10 -- # set +x 00:26:16.724 ************************************ 00:26:16.724 START TEST bdev_raid 00:26:16.724 ************************************ 00:26:16.724 13:46:23 bdev_raid -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh 00:26:16.724 * Looking for test storage... 00:26:16.724 * Found test storage at /home/vagrant/spdk_repo/spdk/test/bdev 00:26:16.724 13:46:23 bdev_raid -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:26:16.724 13:46:23 bdev_raid -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:26:16.724 13:46:23 bdev_raid -- common/autotest_common.sh@1693 -- # lcov --version 00:26:16.724 13:46:23 bdev_raid -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@336 -- # IFS=.-: 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@336 -- # read -ra ver1 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@337 -- # IFS=.-: 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@337 -- # read -ra ver2 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@338 -- # local 'op=<' 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@340 -- # ver1_l=2 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@341 -- # ver2_l=1 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@344 -- # case "$op" in 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@345 -- # : 1 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@364 -- # (( v = 0 )) 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@365 -- # decimal 1 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@353 -- # local d=1 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@355 -- # echo 1 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@365 -- # ver1[v]=1 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@366 -- # decimal 2 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@353 -- # local d=2 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@355 -- # echo 2 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@366 -- # ver2[v]=2 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:26:16.724 13:46:23 bdev_raid -- scripts/common.sh@368 -- # return 0 00:26:16.724 13:46:23 bdev_raid -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:26:16.724 13:46:23 bdev_raid -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:26:16.724 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:16.724 --rc genhtml_branch_coverage=1 00:26:16.724 --rc genhtml_function_coverage=1 00:26:16.724 --rc genhtml_legend=1 00:26:16.724 --rc geninfo_all_blocks=1 00:26:16.724 --rc geninfo_unexecuted_blocks=1 00:26:16.724 00:26:16.724 ' 00:26:16.724 13:46:23 bdev_raid -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:26:16.724 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:16.724 --rc genhtml_branch_coverage=1 00:26:16.724 --rc genhtml_function_coverage=1 00:26:16.725 --rc genhtml_legend=1 00:26:16.725 --rc geninfo_all_blocks=1 00:26:16.725 --rc geninfo_unexecuted_blocks=1 00:26:16.725 00:26:16.725 ' 00:26:16.725 13:46:23 bdev_raid -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:26:16.725 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:16.725 --rc genhtml_branch_coverage=1 00:26:16.725 --rc genhtml_function_coverage=1 00:26:16.725 --rc genhtml_legend=1 00:26:16.725 --rc geninfo_all_blocks=1 00:26:16.725 --rc geninfo_unexecuted_blocks=1 00:26:16.725 00:26:16.725 ' 00:26:16.725 13:46:23 bdev_raid -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:26:16.725 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:16.725 --rc genhtml_branch_coverage=1 00:26:16.725 --rc genhtml_function_coverage=1 00:26:16.725 --rc genhtml_legend=1 00:26:16.725 --rc geninfo_all_blocks=1 00:26:16.725 --rc geninfo_unexecuted_blocks=1 00:26:16.725 00:26:16.725 ' 00:26:16.725 13:46:23 bdev_raid -- bdev/bdev_raid.sh@12 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:26:16.725 13:46:23 bdev_raid -- bdev/nbd_common.sh@6 -- # set -e 00:26:16.725 13:46:23 bdev_raid -- bdev/bdev_raid.sh@14 -- # rpc_py=rpc_cmd 00:26:16.725 13:46:23 bdev_raid -- bdev/bdev_raid.sh@946 -- # mkdir -p /raidtest 00:26:16.725 13:46:23 bdev_raid -- bdev/bdev_raid.sh@947 -- # trap 'cleanup; exit 1' EXIT 00:26:16.725 13:46:23 bdev_raid -- bdev/bdev_raid.sh@949 -- # base_blocklen=512 00:26:16.725 13:46:23 bdev_raid -- bdev/bdev_raid.sh@951 -- # run_test raid1_resize_data_offset_test raid_resize_data_offset_test 00:26:16.725 13:46:23 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:26:16.725 13:46:23 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:26:16.725 13:46:23 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:26:16.725 ************************************ 00:26:16.725 START TEST raid1_resize_data_offset_test 00:26:16.725 ************************************ 00:26:16.725 Process raid pid: 58824 00:26:16.725 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:16.725 13:46:23 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@1129 -- # raid_resize_data_offset_test 00:26:16.725 13:46:23 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@917 -- # raid_pid=58824 00:26:16.725 13:46:23 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@918 -- # echo 'Process raid pid: 58824' 00:26:16.725 13:46:23 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@919 -- # waitforlisten 58824 00:26:16.725 13:46:23 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@916 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:26:16.725 13:46:23 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@835 -- # '[' -z 58824 ']' 00:26:16.725 13:46:23 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:16.725 13:46:23 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:16.725 13:46:23 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:16.725 13:46:23 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:16.725 13:46:23 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:26:16.725 [2024-11-20 13:46:23.760260] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:26:16.725 [2024-11-20 13:46:23.760951] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:16.983 [2024-11-20 13:46:23.914890] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:16.983 [2024-11-20 13:46:24.019176] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:17.241 [2024-11-20 13:46:24.144091] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:17.241 [2024-11-20 13:46:24.144145] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:17.807 13:46:24 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:17.807 13:46:24 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@868 -- # return 0 00:26:17.807 13:46:24 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@922 -- # rpc_cmd bdev_malloc_create -b malloc0 64 512 -o 16 00:26:17.807 13:46:24 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:17.807 13:46:24 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:26:17.807 malloc0 00:26:17.807 13:46:24 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:17.807 13:46:24 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@923 -- # rpc_cmd bdev_malloc_create -b malloc1 64 512 -o 16 00:26:17.807 13:46:24 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:17.807 13:46:24 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:26:17.807 malloc1 00:26:17.807 13:46:24 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:17.807 13:46:24 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@924 -- # rpc_cmd bdev_null_create null0 64 512 00:26:17.807 13:46:24 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:17.807 13:46:24 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:26:17.807 null0 00:26:17.807 13:46:24 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:17.808 13:46:24 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@926 -- # rpc_cmd bdev_raid_create -n Raid -r 1 -b ''\''malloc0 malloc1 null0'\''' -s 00:26:17.808 13:46:24 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:17.808 13:46:24 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:26:17.808 [2024-11-20 13:46:24.749379] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc0 is claimed 00:26:17.808 [2024-11-20 13:46:24.751260] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:26:17.808 [2024-11-20 13:46:24.751406] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev null0 is claimed 00:26:17.808 [2024-11-20 13:46:24.751570] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:26:17.808 [2024-11-20 13:46:24.751704] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 129024, blocklen 512 00:26:17.808 [2024-11-20 13:46:24.751993] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:26:17.808 [2024-11-20 13:46:24.752161] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:26:17.808 [2024-11-20 13:46:24.752190] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007780 00:26:17.808 [2024-11-20 13:46:24.752406] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:17.808 13:46:24 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:17.808 13:46:24 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:17.808 13:46:24 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # jq -r '.[].base_bdevs_list[2].data_offset' 00:26:17.808 13:46:24 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:17.808 13:46:24 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:26:17.808 13:46:24 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:17.808 13:46:24 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # (( 2048 == 2048 )) 00:26:17.808 13:46:24 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@931 -- # rpc_cmd bdev_null_delete null0 00:26:17.808 13:46:24 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:17.808 13:46:24 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:26:17.808 [2024-11-20 13:46:24.789393] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: null0 00:26:17.808 13:46:24 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:17.808 13:46:24 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@935 -- # rpc_cmd bdev_malloc_create -b malloc2 512 512 -o 30 00:26:17.808 13:46:24 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:17.808 13:46:24 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:26:18.067 malloc2 00:26:18.067 13:46:25 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:18.067 13:46:25 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@936 -- # rpc_cmd bdev_raid_add_base_bdev Raid malloc2 00:26:18.067 13:46:25 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:18.067 13:46:25 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:26:18.067 [2024-11-20 13:46:25.122388] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:26:18.325 [2024-11-20 13:46:25.132923] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:26:18.325 13:46:25 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:18.325 [2024-11-20 13:46:25.134627] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev Raid 00:26:18.325 13:46:25 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # jq -r '.[].base_bdevs_list[2].data_offset' 00:26:18.325 13:46:25 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:18.325 13:46:25 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:18.325 13:46:25 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:26:18.325 13:46:25 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:18.325 13:46:25 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # (( 2070 == 2070 )) 00:26:18.325 13:46:25 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@941 -- # killprocess 58824 00:26:18.325 13:46:25 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@954 -- # '[' -z 58824 ']' 00:26:18.325 13:46:25 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@958 -- # kill -0 58824 00:26:18.325 13:46:25 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@959 -- # uname 00:26:18.325 13:46:25 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:18.325 13:46:25 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 58824 00:26:18.325 killing process with pid 58824 00:26:18.325 13:46:25 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:26:18.325 13:46:25 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:26:18.325 13:46:25 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 58824' 00:26:18.325 13:46:25 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@973 -- # kill 58824 00:26:18.325 13:46:25 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@978 -- # wait 58824 00:26:18.325 [2024-11-20 13:46:25.184860] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:26:18.325 [2024-11-20 13:46:25.185791] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev Raid: Operation canceled 00:26:18.325 [2024-11-20 13:46:25.185948] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:18.325 [2024-11-20 13:46:25.185966] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: malloc2 00:26:18.325 [2024-11-20 13:46:25.205807] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:18.325 [2024-11-20 13:46:25.206115] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:26:18.325 [2024-11-20 13:46:25.206129] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Raid, state offline 00:26:19.268 [2024-11-20 13:46:26.151001] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:26:19.838 ************************************ 00:26:19.838 END TEST raid1_resize_data_offset_test 00:26:19.838 ************************************ 00:26:19.838 13:46:26 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@943 -- # return 0 00:26:19.838 00:26:19.838 real 0m3.087s 00:26:19.838 user 0m3.024s 00:26:19.838 sys 0m0.435s 00:26:19.838 13:46:26 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:26:19.838 13:46:26 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:26:19.838 13:46:26 bdev_raid -- bdev/bdev_raid.sh@953 -- # run_test raid0_resize_superblock_test raid_resize_superblock_test 0 00:26:19.838 13:46:26 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:26:19.838 13:46:26 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:26:19.838 13:46:26 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:26:19.838 ************************************ 00:26:19.838 START TEST raid0_resize_superblock_test 00:26:19.838 ************************************ 00:26:19.838 13:46:26 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@1129 -- # raid_resize_superblock_test 0 00:26:19.838 13:46:26 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@854 -- # local raid_level=0 00:26:19.838 13:46:26 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@857 -- # raid_pid=58891 00:26:19.838 13:46:26 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@858 -- # echo 'Process raid pid: 58891' 00:26:19.838 Process raid pid: 58891 00:26:19.838 13:46:26 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@859 -- # waitforlisten 58891 00:26:19.838 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:19.838 13:46:26 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 58891 ']' 00:26:19.838 13:46:26 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:19.838 13:46:26 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:19.838 13:46:26 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:19.838 13:46:26 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@856 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:26:19.838 13:46:26 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:19.838 13:46:26 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:19.838 [2024-11-20 13:46:26.892130] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:26:19.839 [2024-11-20 13:46:26.892410] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:20.097 [2024-11-20 13:46:27.046103] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:20.097 [2024-11-20 13:46:27.150735] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:20.355 [2024-11-20 13:46:27.275272] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:20.355 [2024-11-20 13:46:27.275492] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:20.921 13:46:27 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:20.921 13:46:27 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:26:20.921 13:46:27 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@861 -- # rpc_cmd bdev_malloc_create -b malloc0 512 512 00:26:20.921 13:46:27 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:20.921 13:46:27 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:21.181 malloc0 00:26:21.181 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:21.181 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@863 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:26:21.181 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:21.181 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:21.181 [2024-11-20 13:46:28.148871] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:26:21.181 [2024-11-20 13:46:28.148948] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:21.181 [2024-11-20 13:46:28.148973] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:26:21.181 [2024-11-20 13:46:28.148993] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:21.181 [2024-11-20 13:46:28.151019] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:21.181 [2024-11-20 13:46:28.151054] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:26:21.181 pt0 00:26:21.181 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:21.181 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@864 -- # rpc_cmd bdev_lvol_create_lvstore pt0 lvs0 00:26:21.181 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:21.181 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:21.181 c14082d7-b286-4ee1-8ceb-903a30645fee 00:26:21.181 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:21.181 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@866 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol0 64 00:26:21.181 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:21.181 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:21.440 cf8270eb-f609-4a45-aa63-09952ee6d22d 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@867 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol1 64 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:21.440 8d3eea6b-dcf4-4317-8318-34d26541da85 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@869 -- # case $raid_level in 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@870 -- # rpc_cmd bdev_raid_create -n Raid -r 0 -z 64 -b ''\''lvs0/lvol0 lvs0/lvol1'\''' -s 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:21.440 [2024-11-20 13:46:28.250275] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev cf8270eb-f609-4a45-aa63-09952ee6d22d is claimed 00:26:21.440 [2024-11-20 13:46:28.250375] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 8d3eea6b-dcf4-4317-8318-34d26541da85 is claimed 00:26:21.440 [2024-11-20 13:46:28.250486] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:26:21.440 [2024-11-20 13:46:28.250500] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 245760, blocklen 512 00:26:21.440 [2024-11-20 13:46:28.250725] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:26:21.440 [2024-11-20 13:46:28.250871] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:26:21.440 [2024-11-20 13:46:28.250880] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007780 00:26:21.440 [2024-11-20 13:46:28.251024] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # jq '.[].num_blocks' 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # (( 64 == 64 )) 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # jq '.[].num_blocks' 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # (( 64 == 64 )) 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # rpc_cmd bdev_get_bdevs -b Raid 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # jq '.[].num_blocks' 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:21.440 [2024-11-20 13:46:28.326537] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # (( 245760 == 245760 )) 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@885 -- # rpc_cmd bdev_lvol_resize lvs0/lvol0 100 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:21.440 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:21.441 [2024-11-20 13:46:28.354514] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:26:21.441 [2024-11-20 13:46:28.354543] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'cf8270eb-f609-4a45-aa63-09952ee6d22d' was resized: old size 131072, new size 204800 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@886 -- # rpc_cmd bdev_lvol_resize lvs0/lvol1 100 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:21.441 [2024-11-20 13:46:28.362391] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:26:21.441 [2024-11-20 13:46:28.362410] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '8d3eea6b-dcf4-4317-8318-34d26541da85' was resized: old size 131072, new size 204800 00:26:21.441 [2024-11-20 13:46:28.362435] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 245760 to 393216 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # jq '.[].num_blocks' 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # (( 100 == 100 )) 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # jq '.[].num_blocks' 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # (( 100 == 100 )) 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # jq '.[].num_blocks' 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # rpc_cmd bdev_get_bdevs -b Raid 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:21.441 [2024-11-20 13:46:28.438480] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # (( 393216 == 393216 )) 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@898 -- # rpc_cmd bdev_passthru_delete pt0 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:21.441 [2024-11-20 13:46:28.470308] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev pt0 being removed: closing lvstore lvs0 00:26:21.441 [2024-11-20 13:46:28.470480] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol0 00:26:21.441 [2024-11-20 13:46:28.470537] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:26:21.441 [2024-11-20 13:46:28.470581] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol1 00:26:21.441 [2024-11-20 13:46:28.470703] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:21.441 [2024-11-20 13:46:28.470750] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:26:21.441 [2024-11-20 13:46:28.470782] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Raid, state offline 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@899 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:21.441 [2024-11-20 13:46:28.478254] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:26:21.441 [2024-11-20 13:46:28.478294] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:21.441 [2024-11-20 13:46:28.478310] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008180 00:26:21.441 [2024-11-20 13:46:28.478320] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:21.441 [2024-11-20 13:46:28.480310] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:21.441 [2024-11-20 13:46:28.480341] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:26:21.441 [2024-11-20 13:46:28.481727] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev cf8270eb-f609-4a45-aa63-09952ee6d22d 00:26:21.441 [2024-11-20 13:46:28.481778] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev cf8270eb-f609-4a45-aa63-09952ee6d22d is claimed 00:26:21.441 [2024-11-20 13:46:28.481861] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 8d3eea6b-dcf4-4317-8318-34d26541da85 00:26:21.441 [2024-11-20 13:46:28.481876] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 8d3eea6b-dcf4-4317-8318-34d26541da85 is claimed 00:26:21.441 [2024-11-20 13:46:28.481995] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev 8d3eea6b-dcf4-4317-8318-34d26541da85 (2) smaller than existing raid bdev Raid (3) 00:26:21.441 [2024-11-20 13:46:28.482015] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev cf8270eb-f609-4a45-aa63-09952ee6d22d: File exists 00:26:21.441 [2024-11-20 13:46:28.482048] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:26:21.441 [2024-11-20 13:46:28.482059] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 393216, blocklen 512 00:26:21.441 [2024-11-20 13:46:28.482269] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:26:21.441 [2024-11-20 13:46:28.482396] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:26:21.441 [2024-11-20 13:46:28.482403] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007b00 00:26:21.441 [2024-11-20 13:46:28.482554] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:21.441 pt0 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@900 -- # rpc_cmd bdev_wait_for_examine 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # rpc_cmd bdev_get_bdevs -b Raid 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:26:21.441 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # jq '.[].num_blocks' 00:26:21.701 [2024-11-20 13:46:28.498702] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:21.701 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:21.701 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:26:21.701 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:26:21.701 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # (( 393216 == 393216 )) 00:26:21.701 13:46:28 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@909 -- # killprocess 58891 00:26:21.701 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 58891 ']' 00:26:21.701 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@958 -- # kill -0 58891 00:26:21.701 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@959 -- # uname 00:26:21.701 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:21.701 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 58891 00:26:21.701 killing process with pid 58891 00:26:21.701 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:26:21.701 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:26:21.701 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 58891' 00:26:21.701 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@973 -- # kill 58891 00:26:21.701 [2024-11-20 13:46:28.555227] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:26:21.701 [2024-11-20 13:46:28.555292] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:21.701 13:46:28 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@978 -- # wait 58891 00:26:21.701 [2024-11-20 13:46:28.555351] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:26:21.701 [2024-11-20 13:46:28.555361] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Raid, state offline 00:26:22.269 [2024-11-20 13:46:29.308489] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:26:23.204 ************************************ 00:26:23.204 END TEST raid0_resize_superblock_test 00:26:23.204 ************************************ 00:26:23.204 13:46:29 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@911 -- # return 0 00:26:23.205 00:26:23.205 real 0m3.088s 00:26:23.205 user 0m3.308s 00:26:23.205 sys 0m0.452s 00:26:23.205 13:46:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:26:23.205 13:46:29 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:23.205 13:46:29 bdev_raid -- bdev/bdev_raid.sh@954 -- # run_test raid1_resize_superblock_test raid_resize_superblock_test 1 00:26:23.205 13:46:29 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:26:23.205 13:46:29 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:26:23.205 13:46:29 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:26:23.205 ************************************ 00:26:23.205 START TEST raid1_resize_superblock_test 00:26:23.205 ************************************ 00:26:23.205 13:46:29 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@1129 -- # raid_resize_superblock_test 1 00:26:23.205 13:46:29 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@854 -- # local raid_level=1 00:26:23.205 13:46:29 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@857 -- # raid_pid=58973 00:26:23.205 Process raid pid: 58973 00:26:23.205 13:46:29 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@858 -- # echo 'Process raid pid: 58973' 00:26:23.205 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:23.205 13:46:29 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@859 -- # waitforlisten 58973 00:26:23.205 13:46:29 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 58973 ']' 00:26:23.205 13:46:29 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:23.205 13:46:29 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:23.205 13:46:29 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@856 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:26:23.205 13:46:29 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:23.205 13:46:29 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:23.205 13:46:29 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:23.205 [2024-11-20 13:46:30.028903] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:26:23.205 [2024-11-20 13:46:30.029197] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:23.205 [2024-11-20 13:46:30.182883] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:23.463 [2024-11-20 13:46:30.287947] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:23.463 [2024-11-20 13:46:30.411656] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:23.463 [2024-11-20 13:46:30.411693] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:24.029 13:46:30 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:24.029 13:46:30 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:26:24.029 13:46:30 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@861 -- # rpc_cmd bdev_malloc_create -b malloc0 512 512 00:26:24.029 13:46:30 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:24.029 13:46:30 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:24.288 malloc0 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@863 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:24.288 [2024-11-20 13:46:31.213611] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:26:24.288 [2024-11-20 13:46:31.213674] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:24.288 [2024-11-20 13:46:31.213694] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:26:24.288 [2024-11-20 13:46:31.213705] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:24.288 [2024-11-20 13:46:31.215583] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:24.288 [2024-11-20 13:46:31.215615] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:26:24.288 pt0 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@864 -- # rpc_cmd bdev_lvol_create_lvstore pt0 lvs0 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:24.288 c4af69dd-a0c8-482e-836e-2662e2473c0f 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@866 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol0 64 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:24.288 9783342a-a147-49a6-8d7d-8cafbd9fe1f5 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@867 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol1 64 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:24.288 5b6d0349-220f-4b74-8d87-e82fe341a5ed 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@869 -- # case $raid_level in 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@871 -- # rpc_cmd bdev_raid_create -n Raid -r 1 -b ''\''lvs0/lvol0 lvs0/lvol1'\''' -s 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:24.288 [2024-11-20 13:46:31.313030] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 9783342a-a147-49a6-8d7d-8cafbd9fe1f5 is claimed 00:26:24.288 [2024-11-20 13:46:31.313151] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 5b6d0349-220f-4b74-8d87-e82fe341a5ed is claimed 00:26:24.288 [2024-11-20 13:46:31.313269] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:26:24.288 [2024-11-20 13:46:31.313283] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 122880, blocklen 512 00:26:24.288 [2024-11-20 13:46:31.313504] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:26:24.288 [2024-11-20 13:46:31.313657] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:26:24.288 [2024-11-20 13:46:31.313665] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007780 00:26:24.288 [2024-11-20 13:46:31.313790] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # jq '.[].num_blocks' 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:24.288 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:24.547 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # (( 64 == 64 )) 00:26:24.547 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:26:24.547 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:24.547 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:24.547 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # jq '.[].num_blocks' 00:26:24.547 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:24.547 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # (( 64 == 64 )) 00:26:24.547 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:26:24.547 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # rpc_cmd bdev_get_bdevs -b Raid 00:26:24.547 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:24.547 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:26:24.547 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # jq '.[].num_blocks' 00:26:24.547 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:24.547 [2024-11-20 13:46:31.389259] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:24.547 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:24.547 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:26:24.547 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:26:24.547 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # (( 122880 == 122880 )) 00:26:24.547 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@885 -- # rpc_cmd bdev_lvol_resize lvs0/lvol0 100 00:26:24.547 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:24.547 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:24.547 [2024-11-20 13:46:31.417232] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:26:24.547 [2024-11-20 13:46:31.417255] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '9783342a-a147-49a6-8d7d-8cafbd9fe1f5' was resized: old size 131072, new size 204800 00:26:24.547 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:24.547 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@886 -- # rpc_cmd bdev_lvol_resize lvs0/lvol1 100 00:26:24.547 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:24.548 [2024-11-20 13:46:31.425141] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:26:24.548 [2024-11-20 13:46:31.425159] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '5b6d0349-220f-4b74-8d87-e82fe341a5ed' was resized: old size 131072, new size 204800 00:26:24.548 [2024-11-20 13:46:31.425181] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 122880 to 196608 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # jq '.[].num_blocks' 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # (( 100 == 100 )) 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # jq '.[].num_blocks' 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # (( 100 == 100 )) 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # rpc_cmd bdev_get_bdevs -b Raid 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # jq '.[].num_blocks' 00:26:24.548 [2024-11-20 13:46:31.505278] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # (( 196608 == 196608 )) 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@898 -- # rpc_cmd bdev_passthru_delete pt0 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:24.548 [2024-11-20 13:46:31.537096] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev pt0 being removed: closing lvstore lvs0 00:26:24.548 [2024-11-20 13:46:31.537164] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol0 00:26:24.548 [2024-11-20 13:46:31.537191] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol1 00:26:24.548 [2024-11-20 13:46:31.537336] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:26:24.548 [2024-11-20 13:46:31.537511] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:24.548 [2024-11-20 13:46:31.537566] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:26:24.548 [2024-11-20 13:46:31.537578] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Raid, state offline 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@899 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:24.548 [2024-11-20 13:46:31.545023] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:26:24.548 [2024-11-20 13:46:31.545064] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:24.548 [2024-11-20 13:46:31.545081] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008180 00:26:24.548 [2024-11-20 13:46:31.545093] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:24.548 [2024-11-20 13:46:31.547060] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:24.548 [2024-11-20 13:46:31.547089] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:26:24.548 [2024-11-20 13:46:31.548491] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 9783342a-a147-49a6-8d7d-8cafbd9fe1f5 00:26:24.548 [2024-11-20 13:46:31.548649] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 9783342a-a147-49a6-8d7d-8cafbd9fe1f5 is claimed 00:26:24.548 [2024-11-20 13:46:31.548746] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 5b6d0349-220f-4b74-8d87-e82fe341a5ed 00:26:24.548 [2024-11-20 13:46:31.548763] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 5b6d0349-220f-4b74-8d87-e82fe341a5ed is claimed 00:26:24.548 [2024-11-20 13:46:31.548864] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev 5b6d0349-220f-4b74-8d87-e82fe341a5ed (2) smaller than existing raid bdev Raid (3) 00:26:24.548 [2024-11-20 13:46:31.548882] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev 9783342a-a147-49a6-8d7d-8cafbd9fe1f5: File exists 00:26:24.548 [2024-11-20 13:46:31.548918] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:26:24.548 [2024-11-20 13:46:31.548928] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:26:24.548 [2024-11-20 13:46:31.549150] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:26:24.548 [2024-11-20 13:46:31.549279] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:26:24.548 [2024-11-20 13:46:31.549286] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007b00 00:26:24.548 [2024-11-20 13:46:31.549441] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:24.548 pt0 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@900 -- # rpc_cmd bdev_wait_for_examine 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # rpc_cmd bdev_get_bdevs -b Raid 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # jq '.[].num_blocks' 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:24.548 [2024-11-20 13:46:31.565593] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # (( 196608 == 196608 )) 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@909 -- # killprocess 58973 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 58973 ']' 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@958 -- # kill -0 58973 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@959 -- # uname 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:24.548 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 58973 00:26:24.806 killing process with pid 58973 00:26:24.806 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:26:24.806 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:26:24.806 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 58973' 00:26:24.806 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@973 -- # kill 58973 00:26:24.806 [2024-11-20 13:46:31.616327] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:26:24.806 13:46:31 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@978 -- # wait 58973 00:26:24.806 [2024-11-20 13:46:31.616403] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:24.806 [2024-11-20 13:46:31.616455] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:26:24.806 [2024-11-20 13:46:31.616464] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Raid, state offline 00:26:25.371 [2024-11-20 13:46:32.350289] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:26:25.938 ************************************ 00:26:25.938 END TEST raid1_resize_superblock_test 00:26:25.938 ************************************ 00:26:25.938 13:46:32 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@911 -- # return 0 00:26:25.938 00:26:25.938 real 0m3.009s 00:26:25.938 user 0m3.182s 00:26:25.938 sys 0m0.439s 00:26:25.938 13:46:32 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:26:25.938 13:46:32 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:26.197 13:46:33 bdev_raid -- bdev/bdev_raid.sh@956 -- # uname -s 00:26:26.197 13:46:33 bdev_raid -- bdev/bdev_raid.sh@956 -- # '[' Linux = Linux ']' 00:26:26.197 13:46:33 bdev_raid -- bdev/bdev_raid.sh@956 -- # modprobe -n nbd 00:26:26.197 13:46:33 bdev_raid -- bdev/bdev_raid.sh@957 -- # has_nbd=true 00:26:26.197 13:46:33 bdev_raid -- bdev/bdev_raid.sh@958 -- # modprobe nbd 00:26:26.197 13:46:33 bdev_raid -- bdev/bdev_raid.sh@959 -- # run_test raid_function_test_raid0 raid_function_test raid0 00:26:26.197 13:46:33 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:26:26.197 13:46:33 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:26:26.197 13:46:33 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:26:26.197 ************************************ 00:26:26.197 START TEST raid_function_test_raid0 00:26:26.197 ************************************ 00:26:26.197 13:46:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@1129 -- # raid_function_test raid0 00:26:26.197 13:46:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@64 -- # local raid_level=raid0 00:26:26.197 13:46:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@65 -- # local nbd=/dev/nbd0 00:26:26.197 13:46:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@66 -- # local raid_bdev 00:26:26.197 Process raid pid: 59059 00:26:26.197 13:46:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@69 -- # raid_pid=59059 00:26:26.197 13:46:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@70 -- # echo 'Process raid pid: 59059' 00:26:26.197 13:46:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@71 -- # waitforlisten 59059 00:26:26.197 13:46:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@835 -- # '[' -z 59059 ']' 00:26:26.197 13:46:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:26.197 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:26.197 13:46:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:26.197 13:46:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:26.197 13:46:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:26.197 13:46:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:26:26.197 13:46:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@68 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:26:26.197 [2024-11-20 13:46:33.110821] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:26:26.197 [2024-11-20 13:46:33.111064] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:26.454 [2024-11-20 13:46:33.271171] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:26.454 [2024-11-20 13:46:33.391360] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:26.713 [2024-11-20 13:46:33.541531] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:26.713 [2024-11-20 13:46:33.541583] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:26.972 13:46:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:26.972 13:46:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@868 -- # return 0 00:26:26.972 13:46:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@73 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_1 00:26:26.972 13:46:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:26.972 13:46:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:26:26.972 Base_1 00:26:26.972 13:46:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:26.972 13:46:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@74 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_2 00:26:26.972 13:46:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:26.972 13:46:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:26:26.972 Base_2 00:26:26.972 13:46:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:26.972 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@75 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''Base_1 Base_2'\''' -n raid 00:26:26.972 13:46:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:26.972 13:46:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:26:26.972 [2024-11-20 13:46:34.006857] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:26:26.972 [2024-11-20 13:46:34.009033] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:26:26.972 [2024-11-20 13:46:34.009100] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:26:26.972 [2024-11-20 13:46:34.009113] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:26:26.972 [2024-11-20 13:46:34.009384] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:26:26.972 [2024-11-20 13:46:34.009538] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:26:26.972 [2024-11-20 13:46:34.009547] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid, raid_bdev 0x617000007780 00:26:26.972 [2024-11-20 13:46:34.009690] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:26.972 13:46:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:26.972 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # jq -r '.[0]["name"] | select(.)' 00:26:26.972 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # rpc_cmd bdev_raid_get_bdevs online 00:26:26.972 13:46:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:26.972 13:46:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:26:26.972 13:46:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # raid_bdev=raid 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@78 -- # '[' raid = '' ']' 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@83 -- # nbd_start_disks /var/tmp/spdk.sock raid /dev/nbd0 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@10 -- # bdev_list=('raid') 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@10 -- # local bdev_list 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@11 -- # local nbd_list 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@12 -- # local i 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid /dev/nbd0 00:26:27.230 [2024-11-20 13:46:34.198991] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:26:27.230 /dev/nbd0 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@873 -- # local i 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@877 -- # break 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:26:27.230 1+0 records in 00:26:27.230 1+0 records out 00:26:27.230 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000217815 s, 18.8 MB/s 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@890 -- # size=4096 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@893 -- # return 0 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@84 -- # nbd_get_count /var/tmp/spdk.sock 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:26:27.230 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:26:27.487 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:26:27.487 { 00:26:27.487 "nbd_device": "/dev/nbd0", 00:26:27.487 "bdev_name": "raid" 00:26:27.487 } 00:26:27.487 ]' 00:26:27.487 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # echo '[ 00:26:27.487 { 00:26:27.487 "nbd_device": "/dev/nbd0", 00:26:27.487 "bdev_name": "raid" 00:26:27.487 } 00:26:27.487 ]' 00:26:27.487 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:26:27.487 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:26:27.487 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:26:27.487 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:26:27.487 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # count=1 00:26:27.487 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@66 -- # echo 1 00:26:27.487 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@84 -- # count=1 00:26:27.487 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@85 -- # '[' 1 -ne 1 ']' 00:26:27.487 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@89 -- # raid_unmap_data_verify /dev/nbd0 00:26:27.487 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@17 -- # hash blkdiscard 00:26:27.488 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@18 -- # local nbd=/dev/nbd0 00:26:27.488 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@19 -- # local blksize 00:26:27.488 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # grep -v LOG-SEC 00:26:27.488 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # lsblk -o LOG-SEC /dev/nbd0 00:26:27.488 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # cut -d ' ' -f 5 00:26:27.488 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # blksize=512 00:26:27.488 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@21 -- # local rw_blk_num=4096 00:26:27.488 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@22 -- # local rw_len=2097152 00:26:27.488 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@23 -- # unmap_blk_offs=('0' '1028' '321') 00:26:27.488 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@23 -- # local unmap_blk_offs 00:26:27.488 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@24 -- # unmap_blk_nums=('128' '2035' '456') 00:26:27.488 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@24 -- # local unmap_blk_nums 00:26:27.488 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@25 -- # local unmap_off 00:26:27.488 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@26 -- # local unmap_len 00:26:27.488 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@29 -- # dd if=/dev/urandom of=/raidtest/raidrandtest bs=512 count=4096 00:26:27.488 4096+0 records in 00:26:27.488 4096+0 records out 00:26:27.488 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.0175654 s, 119 MB/s 00:26:27.488 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@30 -- # dd if=/raidtest/raidrandtest of=/dev/nbd0 bs=512 count=4096 oflag=direct 00:26:27.745 4096+0 records in 00:26:27.745 4096+0 records out 00:26:27.745 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.269734 s, 7.8 MB/s 00:26:27.745 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@31 -- # blockdev --flushbufs /dev/nbd0 00:26:27.745 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@34 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:26:28.023 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i = 0 )) 00:26:28.023 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:26:28.023 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=0 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=65536 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=0 count=128 conv=notrunc 00:26:28.024 128+0 records in 00:26:28.024 128+0 records out 00:26:28.024 65536 bytes (66 kB, 64 KiB) copied, 0.000396876 s, 165 MB/s 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 0 -l 65536 /dev/nbd0 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=526336 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=1041920 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=1028 count=2035 conv=notrunc 00:26:28.024 2035+0 records in 00:26:28.024 2035+0 records out 00:26:28.024 1041920 bytes (1.0 MB, 1018 KiB) copied, 0.00890707 s, 117 MB/s 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 526336 -l 1041920 /dev/nbd0 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=164352 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=233472 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=321 count=456 conv=notrunc 00:26:28.024 456+0 records in 00:26:28.024 456+0 records out 00:26:28.024 233472 bytes (233 kB, 228 KiB) copied, 0.00173513 s, 135 MB/s 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 164352 -l 233472 /dev/nbd0 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@52 -- # return 0 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@91 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@50 -- # local nbd_list 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@51 -- # local i 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:26:28.024 13:46:34 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:26:28.281 13:46:35 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:26:28.281 [2024-11-20 13:46:35.085476] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:28.281 13:46:35 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:26:28.281 13:46:35 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:26:28.281 13:46:35 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:26:28.281 13:46:35 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:26:28.281 13:46:35 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:26:28.281 13:46:35 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@41 -- # break 00:26:28.281 13:46:35 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@45 -- # return 0 00:26:28.281 13:46:35 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@92 -- # nbd_get_count /var/tmp/spdk.sock 00:26:28.281 13:46:35 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:26:28.281 13:46:35 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:26:28.538 13:46:35 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:26:28.538 13:46:35 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # echo '[]' 00:26:28.538 13:46:35 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:26:28.538 13:46:35 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:26:28.538 13:46:35 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # echo '' 00:26:28.538 13:46:35 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:26:28.538 13:46:35 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # true 00:26:28.538 13:46:35 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # count=0 00:26:28.538 13:46:35 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@66 -- # echo 0 00:26:28.538 13:46:35 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@92 -- # count=0 00:26:28.538 13:46:35 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@93 -- # '[' 0 -ne 0 ']' 00:26:28.538 13:46:35 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@97 -- # killprocess 59059 00:26:28.538 13:46:35 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@954 -- # '[' -z 59059 ']' 00:26:28.538 13:46:35 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@958 -- # kill -0 59059 00:26:28.538 13:46:35 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@959 -- # uname 00:26:28.538 13:46:35 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:28.538 13:46:35 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 59059 00:26:28.538 killing process with pid 59059 00:26:28.538 13:46:35 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:26:28.538 13:46:35 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:26:28.538 13:46:35 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@972 -- # echo 'killing process with pid 59059' 00:26:28.538 13:46:35 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@973 -- # kill 59059 00:26:28.538 [2024-11-20 13:46:35.420932] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:26:28.538 13:46:35 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@978 -- # wait 59059 00:26:28.538 [2024-11-20 13:46:35.421061] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:28.538 [2024-11-20 13:46:35.421116] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:26:28.538 [2024-11-20 13:46:35.421133] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid, state offline 00:26:28.538 [2024-11-20 13:46:35.556209] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:26:29.471 ************************************ 00:26:29.471 END TEST raid_function_test_raid0 00:26:29.471 ************************************ 00:26:29.471 13:46:36 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@99 -- # return 0 00:26:29.471 00:26:29.471 real 0m3.241s 00:26:29.471 user 0m3.866s 00:26:29.471 sys 0m0.759s 00:26:29.471 13:46:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@1130 -- # xtrace_disable 00:26:29.471 13:46:36 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:26:29.471 13:46:36 bdev_raid -- bdev/bdev_raid.sh@960 -- # run_test raid_function_test_concat raid_function_test concat 00:26:29.471 13:46:36 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:26:29.471 13:46:36 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:26:29.471 13:46:36 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:26:29.471 ************************************ 00:26:29.471 START TEST raid_function_test_concat 00:26:29.471 ************************************ 00:26:29.471 13:46:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@1129 -- # raid_function_test concat 00:26:29.471 13:46:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@64 -- # local raid_level=concat 00:26:29.471 13:46:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@65 -- # local nbd=/dev/nbd0 00:26:29.471 13:46:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@66 -- # local raid_bdev 00:26:29.471 13:46:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@69 -- # raid_pid=59177 00:26:29.471 Process raid pid: 59177 00:26:29.471 13:46:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@70 -- # echo 'Process raid pid: 59177' 00:26:29.471 13:46:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@71 -- # waitforlisten 59177 00:26:29.471 13:46:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@835 -- # '[' -z 59177 ']' 00:26:29.471 13:46:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:29.471 13:46:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:29.471 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:29.471 13:46:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:29.471 13:46:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:29.471 13:46:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:26:29.471 13:46:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@68 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:26:29.471 [2024-11-20 13:46:36.389762] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:26:29.471 [2024-11-20 13:46:36.389896] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:29.729 [2024-11-20 13:46:36.540017] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:29.729 [2024-11-20 13:46:36.646102] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:29.729 [2024-11-20 13:46:36.772279] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:29.729 [2024-11-20 13:46:36.772320] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@868 -- # return 0 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@73 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_1 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:26:30.295 Base_1 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@74 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_2 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:26:30.295 Base_2 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@75 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''Base_1 Base_2'\''' -n raid 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:26:30.295 [2024-11-20 13:46:37.270010] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:26:30.295 [2024-11-20 13:46:37.271688] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:26:30.295 [2024-11-20 13:46:37.271757] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:26:30.295 [2024-11-20 13:46:37.271768] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:26:30.295 [2024-11-20 13:46:37.272027] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:26:30.295 [2024-11-20 13:46:37.272158] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:26:30.295 [2024-11-20 13:46:37.272171] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid, raid_bdev 0x617000007780 00:26:30.295 [2024-11-20 13:46:37.272300] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # rpc_cmd bdev_raid_get_bdevs online 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # jq -r '.[0]["name"] | select(.)' 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # raid_bdev=raid 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@78 -- # '[' raid = '' ']' 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@83 -- # nbd_start_disks /var/tmp/spdk.sock raid /dev/nbd0 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@10 -- # bdev_list=('raid') 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@12 -- # local i 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:26:30.295 13:46:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid /dev/nbd0 00:26:30.603 [2024-11-20 13:46:37.498122] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:26:30.603 /dev/nbd0 00:26:30.603 13:46:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:26:30.603 13:46:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:26:30.603 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:26:30.603 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@873 -- # local i 00:26:30.603 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:26:30.603 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:26:30.603 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:26:30.603 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@877 -- # break 00:26:30.603 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:26:30.603 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:26:30.603 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:26:30.603 1+0 records in 00:26:30.603 1+0 records out 00:26:30.603 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000236418 s, 17.3 MB/s 00:26:30.603 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:26:30.603 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@890 -- # size=4096 00:26:30.603 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:26:30.603 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:26:30.603 13:46:37 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@893 -- # return 0 00:26:30.603 13:46:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:26:30.603 13:46:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:26:30.603 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@84 -- # nbd_get_count /var/tmp/spdk.sock 00:26:30.603 13:46:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:26:30.603 13:46:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:26:30.875 { 00:26:30.875 "nbd_device": "/dev/nbd0", 00:26:30.875 "bdev_name": "raid" 00:26:30.875 } 00:26:30.875 ]' 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # echo '[ 00:26:30.875 { 00:26:30.875 "nbd_device": "/dev/nbd0", 00:26:30.875 "bdev_name": "raid" 00:26:30.875 } 00:26:30.875 ]' 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # count=1 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@66 -- # echo 1 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@84 -- # count=1 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@85 -- # '[' 1 -ne 1 ']' 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@89 -- # raid_unmap_data_verify /dev/nbd0 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@17 -- # hash blkdiscard 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@18 -- # local nbd=/dev/nbd0 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@19 -- # local blksize 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # cut -d ' ' -f 5 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # lsblk -o LOG-SEC /dev/nbd0 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # grep -v LOG-SEC 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # blksize=512 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@21 -- # local rw_blk_num=4096 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@22 -- # local rw_len=2097152 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@23 -- # unmap_blk_offs=('0' '1028' '321') 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@23 -- # local unmap_blk_offs 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@24 -- # unmap_blk_nums=('128' '2035' '456') 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@24 -- # local unmap_blk_nums 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@25 -- # local unmap_off 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@26 -- # local unmap_len 00:26:30.875 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@29 -- # dd if=/dev/urandom of=/raidtest/raidrandtest bs=512 count=4096 00:26:30.875 4096+0 records in 00:26:30.875 4096+0 records out 00:26:30.875 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.0216211 s, 97.0 MB/s 00:26:30.876 13:46:37 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@30 -- # dd if=/raidtest/raidrandtest of=/dev/nbd0 bs=512 count=4096 oflag=direct 00:26:31.134 4096+0 records in 00:26:31.134 4096+0 records out 00:26:31.134 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.242773 s, 8.6 MB/s 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@31 -- # blockdev --flushbufs /dev/nbd0 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@34 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i = 0 )) 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=0 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=65536 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=0 count=128 conv=notrunc 00:26:31.134 128+0 records in 00:26:31.134 128+0 records out 00:26:31.134 65536 bytes (66 kB, 64 KiB) copied, 0.000421292 s, 156 MB/s 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 0 -l 65536 /dev/nbd0 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=526336 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=1041920 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=1028 count=2035 conv=notrunc 00:26:31.134 2035+0 records in 00:26:31.134 2035+0 records out 00:26:31.134 1041920 bytes (1.0 MB, 1018 KiB) copied, 0.00582448 s, 179 MB/s 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 526336 -l 1041920 /dev/nbd0 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=164352 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=233472 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=321 count=456 conv=notrunc 00:26:31.134 456+0 records in 00:26:31.134 456+0 records out 00:26:31.134 233472 bytes (233 kB, 228 KiB) copied, 0.00156055 s, 150 MB/s 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 164352 -l 233472 /dev/nbd0 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@52 -- # return 0 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@91 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@51 -- # local i 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:26:31.134 13:46:38 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:26:31.392 13:46:38 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:26:31.392 13:46:38 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:26:31.392 13:46:38 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:26:31.392 13:46:38 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:26:31.392 13:46:38 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:26:31.392 13:46:38 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:26:31.392 [2024-11-20 13:46:38.380671] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:31.392 13:46:38 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@41 -- # break 00:26:31.392 13:46:38 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@45 -- # return 0 00:26:31.392 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@92 -- # nbd_get_count /var/tmp/spdk.sock 00:26:31.392 13:46:38 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:26:31.392 13:46:38 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:26:31.650 13:46:38 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:26:31.650 13:46:38 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:26:31.650 13:46:38 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:26:31.650 13:46:38 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:26:31.650 13:46:38 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:26:31.650 13:46:38 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # echo '' 00:26:31.650 13:46:38 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # true 00:26:31.650 13:46:38 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # count=0 00:26:31.650 13:46:38 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@66 -- # echo 0 00:26:31.650 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@92 -- # count=0 00:26:31.650 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@93 -- # '[' 0 -ne 0 ']' 00:26:31.650 13:46:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@97 -- # killprocess 59177 00:26:31.650 13:46:38 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@954 -- # '[' -z 59177 ']' 00:26:31.650 13:46:38 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@958 -- # kill -0 59177 00:26:31.650 13:46:38 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@959 -- # uname 00:26:31.650 13:46:38 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:31.650 13:46:38 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 59177 00:26:31.650 13:46:38 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:26:31.650 13:46:38 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:26:31.650 killing process with pid 59177 00:26:31.650 13:46:38 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@972 -- # echo 'killing process with pid 59177' 00:26:31.650 13:46:38 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@973 -- # kill 59177 00:26:31.650 [2024-11-20 13:46:38.625894] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:26:31.650 13:46:38 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@978 -- # wait 59177 00:26:31.650 [2024-11-20 13:46:38.626018] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:31.650 [2024-11-20 13:46:38.626084] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:26:31.650 [2024-11-20 13:46:38.626096] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid, state offline 00:26:31.908 [2024-11-20 13:46:38.734643] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:26:32.474 13:46:39 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@99 -- # return 0 00:26:32.474 00:26:32.474 real 0m3.033s 00:26:32.474 user 0m3.595s 00:26:32.474 sys 0m0.784s 00:26:32.474 ************************************ 00:26:32.474 END TEST raid_function_test_concat 00:26:32.474 ************************************ 00:26:32.474 13:46:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@1130 -- # xtrace_disable 00:26:32.474 13:46:39 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:26:32.474 13:46:39 bdev_raid -- bdev/bdev_raid.sh@963 -- # run_test raid0_resize_test raid_resize_test 0 00:26:32.474 13:46:39 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:26:32.474 13:46:39 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:26:32.474 13:46:39 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:26:32.474 ************************************ 00:26:32.474 START TEST raid0_resize_test 00:26:32.474 ************************************ 00:26:32.474 13:46:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@1129 -- # raid_resize_test 0 00:26:32.474 13:46:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@332 -- # local raid_level=0 00:26:32.474 13:46:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@333 -- # local blksize=512 00:26:32.474 13:46:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@334 -- # local bdev_size_mb=32 00:26:32.474 13:46:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@335 -- # local new_bdev_size_mb=64 00:26:32.474 13:46:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@336 -- # local blkcnt 00:26:32.474 13:46:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@337 -- # local raid_size_mb 00:26:32.474 13:46:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@338 -- # local new_raid_size_mb 00:26:32.474 13:46:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@339 -- # local expected_size 00:26:32.474 13:46:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@342 -- # raid_pid=59294 00:26:32.474 Process raid pid: 59294 00:26:32.474 13:46:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@343 -- # echo 'Process raid pid: 59294' 00:26:32.474 13:46:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@344 -- # waitforlisten 59294 00:26:32.474 13:46:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@835 -- # '[' -z 59294 ']' 00:26:32.474 13:46:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:32.474 13:46:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@341 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:26:32.474 13:46:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:32.474 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:32.474 13:46:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:32.474 13:46:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:32.474 13:46:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:26:32.474 [2024-11-20 13:46:39.453432] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:26:32.474 [2024-11-20 13:46:39.453540] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:32.732 [2024-11-20 13:46:39.604457] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:32.732 [2024-11-20 13:46:39.710644] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:32.989 [2024-11-20 13:46:39.837503] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:32.989 [2024-11-20 13:46:39.837554] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@868 -- # return 0 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@346 -- # rpc_cmd bdev_null_create Base_1 32 512 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:26:33.555 Base_1 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@347 -- # rpc_cmd bdev_null_create Base_2 32 512 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:26:33.555 Base_2 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@349 -- # '[' 0 -eq 0 ']' 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@350 -- # rpc_cmd bdev_raid_create -z 64 -r 0 -b ''\''Base_1 Base_2'\''' -n Raid 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:26:33.555 [2024-11-20 13:46:40.342581] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:26:33.555 [2024-11-20 13:46:40.344364] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:26:33.555 [2024-11-20 13:46:40.344427] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:26:33.555 [2024-11-20 13:46:40.344438] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:26:33.555 [2024-11-20 13:46:40.344704] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:26:33.555 [2024-11-20 13:46:40.344818] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:26:33.555 [2024-11-20 13:46:40.344831] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007780 00:26:33.555 [2024-11-20 13:46:40.344974] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@356 -- # rpc_cmd bdev_null_resize Base_1 64 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:26:33.555 [2024-11-20 13:46:40.350564] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:26:33.555 [2024-11-20 13:46:40.350593] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_1' was resized: old size 65536, new size 131072 00:26:33.555 true 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # rpc_cmd bdev_get_bdevs -b Raid 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # jq '.[].num_blocks' 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:26:33.555 [2024-11-20 13:46:40.362747] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # blkcnt=131072 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@360 -- # raid_size_mb=64 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@361 -- # '[' 0 -eq 0 ']' 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@362 -- # expected_size=64 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@366 -- # '[' 64 '!=' 64 ']' 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@372 -- # rpc_cmd bdev_null_resize Base_2 64 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:26:33.555 [2024-11-20 13:46:40.394603] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:26:33.555 [2024-11-20 13:46:40.394643] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_2' was resized: old size 65536, new size 131072 00:26:33.555 [2024-11-20 13:46:40.394676] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 131072 to 262144 00:26:33.555 true 00:26:33.555 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:33.556 13:46:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # rpc_cmd bdev_get_bdevs -b Raid 00:26:33.556 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:33.556 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:26:33.556 13:46:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # jq '.[].num_blocks' 00:26:33.556 [2024-11-20 13:46:40.406764] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:33.556 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:33.556 13:46:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # blkcnt=262144 00:26:33.556 13:46:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@376 -- # raid_size_mb=128 00:26:33.556 13:46:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@377 -- # '[' 0 -eq 0 ']' 00:26:33.556 13:46:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@378 -- # expected_size=128 00:26:33.556 13:46:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@382 -- # '[' 128 '!=' 128 ']' 00:26:33.556 13:46:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@387 -- # killprocess 59294 00:26:33.556 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@954 -- # '[' -z 59294 ']' 00:26:33.556 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@958 -- # kill -0 59294 00:26:33.556 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@959 -- # uname 00:26:33.556 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:33.556 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 59294 00:26:33.556 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:26:33.556 killing process with pid 59294 00:26:33.556 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:26:33.556 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 59294' 00:26:33.556 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@973 -- # kill 59294 00:26:33.556 13:46:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@978 -- # wait 59294 00:26:33.556 [2024-11-20 13:46:40.463264] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:26:33.556 [2024-11-20 13:46:40.463386] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:33.556 [2024-11-20 13:46:40.463441] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:26:33.556 [2024-11-20 13:46:40.463450] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Raid, state offline 00:26:33.556 [2024-11-20 13:46:40.474177] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:26:34.119 13:46:41 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@389 -- # return 0 00:26:34.119 00:26:34.119 real 0m1.741s 00:26:34.119 user 0m1.888s 00:26:34.119 sys 0m0.281s 00:26:34.119 13:46:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:26:34.119 13:46:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:26:34.119 ************************************ 00:26:34.119 END TEST raid0_resize_test 00:26:34.119 ************************************ 00:26:34.119 13:46:41 bdev_raid -- bdev/bdev_raid.sh@964 -- # run_test raid1_resize_test raid_resize_test 1 00:26:34.119 13:46:41 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:26:34.119 13:46:41 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:26:34.119 13:46:41 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:26:34.377 ************************************ 00:26:34.377 START TEST raid1_resize_test 00:26:34.377 ************************************ 00:26:34.377 13:46:41 bdev_raid.raid1_resize_test -- common/autotest_common.sh@1129 -- # raid_resize_test 1 00:26:34.377 13:46:41 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@332 -- # local raid_level=1 00:26:34.377 13:46:41 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@333 -- # local blksize=512 00:26:34.377 13:46:41 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@334 -- # local bdev_size_mb=32 00:26:34.377 13:46:41 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@335 -- # local new_bdev_size_mb=64 00:26:34.377 13:46:41 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@336 -- # local blkcnt 00:26:34.377 13:46:41 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@337 -- # local raid_size_mb 00:26:34.377 13:46:41 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@338 -- # local new_raid_size_mb 00:26:34.377 13:46:41 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@339 -- # local expected_size 00:26:34.377 Process raid pid: 59350 00:26:34.377 13:46:41 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@342 -- # raid_pid=59350 00:26:34.377 13:46:41 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@343 -- # echo 'Process raid pid: 59350' 00:26:34.377 13:46:41 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@344 -- # waitforlisten 59350 00:26:34.377 13:46:41 bdev_raid.raid1_resize_test -- common/autotest_common.sh@835 -- # '[' -z 59350 ']' 00:26:34.377 13:46:41 bdev_raid.raid1_resize_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:34.377 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:34.377 13:46:41 bdev_raid.raid1_resize_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:34.377 13:46:41 bdev_raid.raid1_resize_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:34.377 13:46:41 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@341 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:26:34.377 13:46:41 bdev_raid.raid1_resize_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:34.377 13:46:41 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:26:34.377 [2024-11-20 13:46:41.253351] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:26:34.377 [2024-11-20 13:46:41.253484] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:34.377 [2024-11-20 13:46:41.413546] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:34.634 [2024-11-20 13:46:41.526173] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:34.634 [2024-11-20 13:46:41.660687] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:34.634 [2024-11-20 13:46:41.660749] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:35.234 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:35.234 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@868 -- # return 0 00:26:35.234 13:46:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@346 -- # rpc_cmd bdev_null_create Base_1 32 512 00:26:35.234 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:35.234 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:26:35.234 Base_1 00:26:35.234 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:35.234 13:46:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@347 -- # rpc_cmd bdev_null_create Base_2 32 512 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:26:35.235 Base_2 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@349 -- # '[' 1 -eq 0 ']' 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@352 -- # rpc_cmd bdev_raid_create -r 1 -b ''\''Base_1 Base_2'\''' -n Raid 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:26:35.235 [2024-11-20 13:46:42.128640] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:26:35.235 [2024-11-20 13:46:42.130300] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:26:35.235 [2024-11-20 13:46:42.130353] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:26:35.235 [2024-11-20 13:46:42.130363] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:26:35.235 [2024-11-20 13:46:42.130592] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:26:35.235 [2024-11-20 13:46:42.130699] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:26:35.235 [2024-11-20 13:46:42.130710] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007780 00:26:35.235 [2024-11-20 13:46:42.130825] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@356 -- # rpc_cmd bdev_null_resize Base_1 64 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:26:35.235 [2024-11-20 13:46:42.136622] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:26:35.235 [2024-11-20 13:46:42.136651] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_1' was resized: old size 65536, new size 131072 00:26:35.235 true 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # rpc_cmd bdev_get_bdevs -b Raid 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # jq '.[].num_blocks' 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:26:35.235 [2024-11-20 13:46:42.148810] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # blkcnt=65536 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@360 -- # raid_size_mb=32 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@361 -- # '[' 1 -eq 0 ']' 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@364 -- # expected_size=32 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@366 -- # '[' 32 '!=' 32 ']' 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@372 -- # rpc_cmd bdev_null_resize Base_2 64 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:26:35.235 [2024-11-20 13:46:42.180638] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:26:35.235 [2024-11-20 13:46:42.180662] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_2' was resized: old size 65536, new size 131072 00:26:35.235 [2024-11-20 13:46:42.180687] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 65536 to 131072 00:26:35.235 true 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # jq '.[].num_blocks' 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # rpc_cmd bdev_get_bdevs -b Raid 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:26:35.235 [2024-11-20 13:46:42.192790] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # blkcnt=131072 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@376 -- # raid_size_mb=64 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@377 -- # '[' 1 -eq 0 ']' 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@380 -- # expected_size=64 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@382 -- # '[' 64 '!=' 64 ']' 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@387 -- # killprocess 59350 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@954 -- # '[' -z 59350 ']' 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@958 -- # kill -0 59350 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@959 -- # uname 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 59350 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:26:35.235 killing process with pid 59350 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 59350' 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@973 -- # kill 59350 00:26:35.235 [2024-11-20 13:46:42.242559] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:26:35.235 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@978 -- # wait 59350 00:26:35.235 [2024-11-20 13:46:42.242651] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:35.235 [2024-11-20 13:46:42.243088] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:26:35.235 [2024-11-20 13:46:42.243109] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Raid, state offline 00:26:35.235 [2024-11-20 13:46:42.252244] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:26:36.218 13:46:42 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@389 -- # return 0 00:26:36.218 00:26:36.218 real 0m1.696s 00:26:36.218 user 0m1.844s 00:26:36.218 sys 0m0.269s 00:26:36.218 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:26:36.218 13:46:42 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:26:36.218 ************************************ 00:26:36.218 END TEST raid1_resize_test 00:26:36.218 ************************************ 00:26:36.218 13:46:42 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:26:36.218 13:46:42 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:26:36.218 13:46:42 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 2 false 00:26:36.218 13:46:42 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:26:36.218 13:46:42 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:26:36.218 13:46:42 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:26:36.218 ************************************ 00:26:36.218 START TEST raid_state_function_test 00:26:36.218 ************************************ 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 2 false 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=59407 00:26:36.218 Process raid pid: 59407 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 59407' 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 59407 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 59407 ']' 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:26:36.218 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:36.218 13:46:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:36.218 [2024-11-20 13:46:43.008662] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:26:36.218 [2024-11-20 13:46:43.008796] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:36.218 [2024-11-20 13:46:43.162114] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:36.478 [2024-11-20 13:46:43.288487] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:36.478 [2024-11-20 13:46:43.441556] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:36.478 [2024-11-20 13:46:43.441603] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:37.052 13:46:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:37.052 13:46:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:26:37.052 13:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:26:37.052 13:46:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:37.052 13:46:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:37.052 [2024-11-20 13:46:43.923169] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:26:37.052 [2024-11-20 13:46:43.923231] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:26:37.052 [2024-11-20 13:46:43.923242] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:26:37.052 [2024-11-20 13:46:43.923252] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:26:37.052 13:46:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:37.052 13:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:26:37.052 13:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:26:37.052 13:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:26:37.052 13:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:26:37.052 13:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:26:37.052 13:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:37.052 13:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:37.052 13:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:37.052 13:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:37.052 13:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:37.052 13:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:26:37.052 13:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:37.052 13:46:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:37.052 13:46:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:37.052 13:46:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:37.052 13:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:37.052 "name": "Existed_Raid", 00:26:37.052 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:37.052 "strip_size_kb": 64, 00:26:37.052 "state": "configuring", 00:26:37.052 "raid_level": "raid0", 00:26:37.052 "superblock": false, 00:26:37.052 "num_base_bdevs": 2, 00:26:37.052 "num_base_bdevs_discovered": 0, 00:26:37.052 "num_base_bdevs_operational": 2, 00:26:37.052 "base_bdevs_list": [ 00:26:37.052 { 00:26:37.052 "name": "BaseBdev1", 00:26:37.052 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:37.052 "is_configured": false, 00:26:37.052 "data_offset": 0, 00:26:37.052 "data_size": 0 00:26:37.052 }, 00:26:37.052 { 00:26:37.052 "name": "BaseBdev2", 00:26:37.052 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:37.052 "is_configured": false, 00:26:37.052 "data_offset": 0, 00:26:37.052 "data_size": 0 00:26:37.052 } 00:26:37.052 ] 00:26:37.052 }' 00:26:37.052 13:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:37.052 13:46:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:37.315 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:26:37.315 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:37.315 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:37.315 [2024-11-20 13:46:44.231251] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:26:37.315 [2024-11-20 13:46:44.231297] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:26:37.315 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:37.315 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:26:37.315 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:37.315 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:37.315 [2024-11-20 13:46:44.243240] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:26:37.315 [2024-11-20 13:46:44.243292] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:26:37.316 [2024-11-20 13:46:44.243302] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:26:37.316 [2024-11-20 13:46:44.243313] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:37.316 [2024-11-20 13:46:44.278719] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:26:37.316 BaseBdev1 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:37.316 [ 00:26:37.316 { 00:26:37.316 "name": "BaseBdev1", 00:26:37.316 "aliases": [ 00:26:37.316 "e2def527-9209-4e05-adb3-5ce5dd35aff6" 00:26:37.316 ], 00:26:37.316 "product_name": "Malloc disk", 00:26:37.316 "block_size": 512, 00:26:37.316 "num_blocks": 65536, 00:26:37.316 "uuid": "e2def527-9209-4e05-adb3-5ce5dd35aff6", 00:26:37.316 "assigned_rate_limits": { 00:26:37.316 "rw_ios_per_sec": 0, 00:26:37.316 "rw_mbytes_per_sec": 0, 00:26:37.316 "r_mbytes_per_sec": 0, 00:26:37.316 "w_mbytes_per_sec": 0 00:26:37.316 }, 00:26:37.316 "claimed": true, 00:26:37.316 "claim_type": "exclusive_write", 00:26:37.316 "zoned": false, 00:26:37.316 "supported_io_types": { 00:26:37.316 "read": true, 00:26:37.316 "write": true, 00:26:37.316 "unmap": true, 00:26:37.316 "flush": true, 00:26:37.316 "reset": true, 00:26:37.316 "nvme_admin": false, 00:26:37.316 "nvme_io": false, 00:26:37.316 "nvme_io_md": false, 00:26:37.316 "write_zeroes": true, 00:26:37.316 "zcopy": true, 00:26:37.316 "get_zone_info": false, 00:26:37.316 "zone_management": false, 00:26:37.316 "zone_append": false, 00:26:37.316 "compare": false, 00:26:37.316 "compare_and_write": false, 00:26:37.316 "abort": true, 00:26:37.316 "seek_hole": false, 00:26:37.316 "seek_data": false, 00:26:37.316 "copy": true, 00:26:37.316 "nvme_iov_md": false 00:26:37.316 }, 00:26:37.316 "memory_domains": [ 00:26:37.316 { 00:26:37.316 "dma_device_id": "system", 00:26:37.316 "dma_device_type": 1 00:26:37.316 }, 00:26:37.316 { 00:26:37.316 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:37.316 "dma_device_type": 2 00:26:37.316 } 00:26:37.316 ], 00:26:37.316 "driver_specific": {} 00:26:37.316 } 00:26:37.316 ] 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:37.316 "name": "Existed_Raid", 00:26:37.316 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:37.316 "strip_size_kb": 64, 00:26:37.316 "state": "configuring", 00:26:37.316 "raid_level": "raid0", 00:26:37.316 "superblock": false, 00:26:37.316 "num_base_bdevs": 2, 00:26:37.316 "num_base_bdevs_discovered": 1, 00:26:37.316 "num_base_bdevs_operational": 2, 00:26:37.316 "base_bdevs_list": [ 00:26:37.316 { 00:26:37.316 "name": "BaseBdev1", 00:26:37.316 "uuid": "e2def527-9209-4e05-adb3-5ce5dd35aff6", 00:26:37.316 "is_configured": true, 00:26:37.316 "data_offset": 0, 00:26:37.316 "data_size": 65536 00:26:37.316 }, 00:26:37.316 { 00:26:37.316 "name": "BaseBdev2", 00:26:37.316 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:37.316 "is_configured": false, 00:26:37.316 "data_offset": 0, 00:26:37.316 "data_size": 0 00:26:37.316 } 00:26:37.316 ] 00:26:37.316 }' 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:37.316 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:37.575 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:26:37.575 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:37.575 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:37.575 [2024-11-20 13:46:44.614848] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:26:37.575 [2024-11-20 13:46:44.615039] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:26:37.575 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:37.575 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:26:37.575 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:37.575 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:37.575 [2024-11-20 13:46:44.622867] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:26:37.575 [2024-11-20 13:46:44.624876] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:26:37.575 [2024-11-20 13:46:44.624915] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:26:37.575 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:37.575 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:26:37.575 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:26:37.575 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:26:37.575 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:26:37.575 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:26:37.575 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:26:37.575 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:26:37.575 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:37.575 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:37.575 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:37.575 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:37.575 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:37.833 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:37.833 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:26:37.833 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:37.833 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:37.833 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:37.833 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:37.833 "name": "Existed_Raid", 00:26:37.833 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:37.833 "strip_size_kb": 64, 00:26:37.833 "state": "configuring", 00:26:37.833 "raid_level": "raid0", 00:26:37.833 "superblock": false, 00:26:37.833 "num_base_bdevs": 2, 00:26:37.833 "num_base_bdevs_discovered": 1, 00:26:37.833 "num_base_bdevs_operational": 2, 00:26:37.833 "base_bdevs_list": [ 00:26:37.833 { 00:26:37.833 "name": "BaseBdev1", 00:26:37.833 "uuid": "e2def527-9209-4e05-adb3-5ce5dd35aff6", 00:26:37.833 "is_configured": true, 00:26:37.833 "data_offset": 0, 00:26:37.833 "data_size": 65536 00:26:37.833 }, 00:26:37.833 { 00:26:37.833 "name": "BaseBdev2", 00:26:37.833 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:37.833 "is_configured": false, 00:26:37.833 "data_offset": 0, 00:26:37.833 "data_size": 0 00:26:37.833 } 00:26:37.833 ] 00:26:37.833 }' 00:26:37.833 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:37.833 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:38.092 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:26:38.092 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:38.092 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:38.092 [2024-11-20 13:46:44.936330] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:26:38.092 [2024-11-20 13:46:44.936555] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:26:38.092 [2024-11-20 13:46:44.936574] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:26:38.092 [2024-11-20 13:46:44.936875] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:26:38.092 [2024-11-20 13:46:44.937059] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:26:38.092 [2024-11-20 13:46:44.937071] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:26:38.092 [2024-11-20 13:46:44.937334] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:38.092 BaseBdev2 00:26:38.092 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:38.092 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:26:38.092 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:26:38.092 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:26:38.092 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:26:38.092 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:26:38.092 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:26:38.092 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:26:38.092 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:38.092 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:38.092 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:38.092 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:26:38.092 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:38.092 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:38.092 [ 00:26:38.092 { 00:26:38.092 "name": "BaseBdev2", 00:26:38.092 "aliases": [ 00:26:38.092 "a1ef070f-a1b0-4752-b8f4-26305be2a23b" 00:26:38.092 ], 00:26:38.092 "product_name": "Malloc disk", 00:26:38.092 "block_size": 512, 00:26:38.092 "num_blocks": 65536, 00:26:38.092 "uuid": "a1ef070f-a1b0-4752-b8f4-26305be2a23b", 00:26:38.092 "assigned_rate_limits": { 00:26:38.092 "rw_ios_per_sec": 0, 00:26:38.092 "rw_mbytes_per_sec": 0, 00:26:38.092 "r_mbytes_per_sec": 0, 00:26:38.092 "w_mbytes_per_sec": 0 00:26:38.092 }, 00:26:38.092 "claimed": true, 00:26:38.092 "claim_type": "exclusive_write", 00:26:38.092 "zoned": false, 00:26:38.092 "supported_io_types": { 00:26:38.092 "read": true, 00:26:38.092 "write": true, 00:26:38.092 "unmap": true, 00:26:38.092 "flush": true, 00:26:38.092 "reset": true, 00:26:38.092 "nvme_admin": false, 00:26:38.093 "nvme_io": false, 00:26:38.093 "nvme_io_md": false, 00:26:38.093 "write_zeroes": true, 00:26:38.093 "zcopy": true, 00:26:38.093 "get_zone_info": false, 00:26:38.093 "zone_management": false, 00:26:38.093 "zone_append": false, 00:26:38.093 "compare": false, 00:26:38.093 "compare_and_write": false, 00:26:38.093 "abort": true, 00:26:38.093 "seek_hole": false, 00:26:38.093 "seek_data": false, 00:26:38.093 "copy": true, 00:26:38.093 "nvme_iov_md": false 00:26:38.093 }, 00:26:38.093 "memory_domains": [ 00:26:38.093 { 00:26:38.093 "dma_device_id": "system", 00:26:38.093 "dma_device_type": 1 00:26:38.093 }, 00:26:38.093 { 00:26:38.093 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:38.093 "dma_device_type": 2 00:26:38.093 } 00:26:38.093 ], 00:26:38.093 "driver_specific": {} 00:26:38.093 } 00:26:38.093 ] 00:26:38.093 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:38.093 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:26:38.093 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:26:38.093 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:26:38.093 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 2 00:26:38.093 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:26:38.093 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:26:38.093 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:26:38.093 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:26:38.093 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:38.093 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:38.093 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:38.093 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:38.093 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:38.093 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:38.093 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:38.093 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:38.093 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:26:38.093 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:38.093 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:38.093 "name": "Existed_Raid", 00:26:38.093 "uuid": "9041045e-34f7-475e-9d3a-7ddbd08fb69a", 00:26:38.093 "strip_size_kb": 64, 00:26:38.093 "state": "online", 00:26:38.093 "raid_level": "raid0", 00:26:38.093 "superblock": false, 00:26:38.093 "num_base_bdevs": 2, 00:26:38.093 "num_base_bdevs_discovered": 2, 00:26:38.093 "num_base_bdevs_operational": 2, 00:26:38.093 "base_bdevs_list": [ 00:26:38.093 { 00:26:38.093 "name": "BaseBdev1", 00:26:38.093 "uuid": "e2def527-9209-4e05-adb3-5ce5dd35aff6", 00:26:38.093 "is_configured": true, 00:26:38.093 "data_offset": 0, 00:26:38.093 "data_size": 65536 00:26:38.093 }, 00:26:38.093 { 00:26:38.093 "name": "BaseBdev2", 00:26:38.093 "uuid": "a1ef070f-a1b0-4752-b8f4-26305be2a23b", 00:26:38.093 "is_configured": true, 00:26:38.093 "data_offset": 0, 00:26:38.093 "data_size": 65536 00:26:38.093 } 00:26:38.093 ] 00:26:38.093 }' 00:26:38.093 13:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:38.093 13:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:38.351 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:26:38.351 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:26:38.351 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:26:38.351 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:26:38.351 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:26:38.351 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:26:38.351 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:26:38.351 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:38.351 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:26:38.351 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:38.351 [2024-11-20 13:46:45.264759] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:38.351 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:38.351 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:26:38.351 "name": "Existed_Raid", 00:26:38.351 "aliases": [ 00:26:38.351 "9041045e-34f7-475e-9d3a-7ddbd08fb69a" 00:26:38.351 ], 00:26:38.351 "product_name": "Raid Volume", 00:26:38.351 "block_size": 512, 00:26:38.351 "num_blocks": 131072, 00:26:38.351 "uuid": "9041045e-34f7-475e-9d3a-7ddbd08fb69a", 00:26:38.351 "assigned_rate_limits": { 00:26:38.351 "rw_ios_per_sec": 0, 00:26:38.351 "rw_mbytes_per_sec": 0, 00:26:38.351 "r_mbytes_per_sec": 0, 00:26:38.351 "w_mbytes_per_sec": 0 00:26:38.351 }, 00:26:38.351 "claimed": false, 00:26:38.351 "zoned": false, 00:26:38.351 "supported_io_types": { 00:26:38.351 "read": true, 00:26:38.351 "write": true, 00:26:38.351 "unmap": true, 00:26:38.351 "flush": true, 00:26:38.351 "reset": true, 00:26:38.351 "nvme_admin": false, 00:26:38.351 "nvme_io": false, 00:26:38.351 "nvme_io_md": false, 00:26:38.351 "write_zeroes": true, 00:26:38.351 "zcopy": false, 00:26:38.351 "get_zone_info": false, 00:26:38.351 "zone_management": false, 00:26:38.351 "zone_append": false, 00:26:38.351 "compare": false, 00:26:38.351 "compare_and_write": false, 00:26:38.351 "abort": false, 00:26:38.351 "seek_hole": false, 00:26:38.351 "seek_data": false, 00:26:38.351 "copy": false, 00:26:38.351 "nvme_iov_md": false 00:26:38.351 }, 00:26:38.351 "memory_domains": [ 00:26:38.351 { 00:26:38.351 "dma_device_id": "system", 00:26:38.351 "dma_device_type": 1 00:26:38.351 }, 00:26:38.351 { 00:26:38.351 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:38.351 "dma_device_type": 2 00:26:38.351 }, 00:26:38.351 { 00:26:38.351 "dma_device_id": "system", 00:26:38.351 "dma_device_type": 1 00:26:38.351 }, 00:26:38.351 { 00:26:38.351 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:38.351 "dma_device_type": 2 00:26:38.351 } 00:26:38.351 ], 00:26:38.351 "driver_specific": { 00:26:38.351 "raid": { 00:26:38.351 "uuid": "9041045e-34f7-475e-9d3a-7ddbd08fb69a", 00:26:38.351 "strip_size_kb": 64, 00:26:38.351 "state": "online", 00:26:38.351 "raid_level": "raid0", 00:26:38.351 "superblock": false, 00:26:38.351 "num_base_bdevs": 2, 00:26:38.351 "num_base_bdevs_discovered": 2, 00:26:38.351 "num_base_bdevs_operational": 2, 00:26:38.351 "base_bdevs_list": [ 00:26:38.351 { 00:26:38.351 "name": "BaseBdev1", 00:26:38.351 "uuid": "e2def527-9209-4e05-adb3-5ce5dd35aff6", 00:26:38.351 "is_configured": true, 00:26:38.351 "data_offset": 0, 00:26:38.351 "data_size": 65536 00:26:38.351 }, 00:26:38.351 { 00:26:38.351 "name": "BaseBdev2", 00:26:38.351 "uuid": "a1ef070f-a1b0-4752-b8f4-26305be2a23b", 00:26:38.351 "is_configured": true, 00:26:38.351 "data_offset": 0, 00:26:38.351 "data_size": 65536 00:26:38.351 } 00:26:38.351 ] 00:26:38.351 } 00:26:38.351 } 00:26:38.351 }' 00:26:38.351 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:26:38.351 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:26:38.351 BaseBdev2' 00:26:38.351 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:26:38.352 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:26:38.352 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:26:38.352 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:26:38.352 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:26:38.352 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:38.352 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:38.352 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:38.352 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:26:38.352 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:26:38.352 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:26:38.352 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:26:38.352 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:26:38.352 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:38.352 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:38.352 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:38.610 [2024-11-20 13:46:45.420558] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:26:38.610 [2024-11-20 13:46:45.420596] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:26:38.610 [2024-11-20 13:46:45.420654] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 1 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:38.610 "name": "Existed_Raid", 00:26:38.610 "uuid": "9041045e-34f7-475e-9d3a-7ddbd08fb69a", 00:26:38.610 "strip_size_kb": 64, 00:26:38.610 "state": "offline", 00:26:38.610 "raid_level": "raid0", 00:26:38.610 "superblock": false, 00:26:38.610 "num_base_bdevs": 2, 00:26:38.610 "num_base_bdevs_discovered": 1, 00:26:38.610 "num_base_bdevs_operational": 1, 00:26:38.610 "base_bdevs_list": [ 00:26:38.610 { 00:26:38.610 "name": null, 00:26:38.610 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:38.610 "is_configured": false, 00:26:38.610 "data_offset": 0, 00:26:38.610 "data_size": 65536 00:26:38.610 }, 00:26:38.610 { 00:26:38.610 "name": "BaseBdev2", 00:26:38.610 "uuid": "a1ef070f-a1b0-4752-b8f4-26305be2a23b", 00:26:38.610 "is_configured": true, 00:26:38.610 "data_offset": 0, 00:26:38.610 "data_size": 65536 00:26:38.610 } 00:26:38.610 ] 00:26:38.610 }' 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:38.610 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:38.869 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:26:38.869 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:26:38.869 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:38.869 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:38.869 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:26:38.869 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:38.869 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:38.869 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:26:38.869 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:26:38.869 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:26:38.869 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:38.869 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:38.869 [2024-11-20 13:46:45.827543] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:26:38.869 [2024-11-20 13:46:45.827715] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:26:38.869 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:38.869 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:26:38.869 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:26:38.869 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:38.869 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:26:38.869 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:38.869 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:38.869 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:39.164 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:26:39.164 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:26:39.164 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:26:39.164 13:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 59407 00:26:39.164 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 59407 ']' 00:26:39.164 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 59407 00:26:39.164 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:26:39.164 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:39.164 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 59407 00:26:39.164 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:26:39.164 killing process with pid 59407 00:26:39.164 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:26:39.164 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 59407' 00:26:39.164 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 59407 00:26:39.164 [2024-11-20 13:46:45.955164] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:26:39.164 13:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 59407 00:26:39.164 [2024-11-20 13:46:45.966410] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:26:39.763 13:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:26:39.763 00:26:39.763 real 0m3.797s 00:26:39.763 user 0m5.395s 00:26:39.763 sys 0m0.600s 00:26:39.763 ************************************ 00:26:39.763 END TEST raid_state_function_test 00:26:39.763 ************************************ 00:26:39.763 13:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:26:39.763 13:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:39.763 13:46:46 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 2 true 00:26:39.763 13:46:46 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:26:39.763 13:46:46 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:26:39.763 13:46:46 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:26:39.763 ************************************ 00:26:39.763 START TEST raid_state_function_test_sb 00:26:39.763 ************************************ 00:26:39.763 13:46:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 2 true 00:26:39.763 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:26:39.763 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:26:39.763 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:26:39.763 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=59643 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 59643' 00:26:39.764 Process raid pid: 59643 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 59643 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 59643 ']' 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:39.764 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:39.764 13:46:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:40.021 [2024-11-20 13:46:46.852266] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:26:40.021 [2024-11-20 13:46:46.852406] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:40.021 [2024-11-20 13:46:47.014680] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:40.278 [2024-11-20 13:46:47.134890] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:40.279 [2024-11-20 13:46:47.284025] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:40.279 [2024-11-20 13:46:47.284078] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:40.845 13:46:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:40.845 13:46:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:26:40.845 13:46:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:26:40.845 13:46:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:40.845 13:46:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:40.845 [2024-11-20 13:46:47.717411] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:26:40.845 [2024-11-20 13:46:47.717471] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:26:40.845 [2024-11-20 13:46:47.717482] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:26:40.845 [2024-11-20 13:46:47.717492] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:26:40.845 13:46:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:40.845 13:46:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:26:40.845 13:46:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:26:40.845 13:46:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:26:40.845 13:46:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:26:40.845 13:46:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:26:40.845 13:46:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:40.845 13:46:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:40.845 13:46:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:40.845 13:46:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:40.845 13:46:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:40.845 13:46:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:26:40.845 13:46:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:40.845 13:46:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:40.845 13:46:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:40.845 13:46:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:40.845 13:46:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:40.845 "name": "Existed_Raid", 00:26:40.845 "uuid": "569ad35a-38ae-4bc1-9f02-b9d194015f27", 00:26:40.845 "strip_size_kb": 64, 00:26:40.845 "state": "configuring", 00:26:40.845 "raid_level": "raid0", 00:26:40.845 "superblock": true, 00:26:40.845 "num_base_bdevs": 2, 00:26:40.845 "num_base_bdevs_discovered": 0, 00:26:40.845 "num_base_bdevs_operational": 2, 00:26:40.845 "base_bdevs_list": [ 00:26:40.845 { 00:26:40.845 "name": "BaseBdev1", 00:26:40.845 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:40.845 "is_configured": false, 00:26:40.845 "data_offset": 0, 00:26:40.845 "data_size": 0 00:26:40.845 }, 00:26:40.845 { 00:26:40.845 "name": "BaseBdev2", 00:26:40.845 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:40.845 "is_configured": false, 00:26:40.845 "data_offset": 0, 00:26:40.845 "data_size": 0 00:26:40.845 } 00:26:40.845 ] 00:26:40.845 }' 00:26:40.845 13:46:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:40.845 13:46:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:41.104 [2024-11-20 13:46:48.033417] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:26:41.104 [2024-11-20 13:46:48.033463] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:41.104 [2024-11-20 13:46:48.041425] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:26:41.104 [2024-11-20 13:46:48.041473] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:26:41.104 [2024-11-20 13:46:48.041483] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:26:41.104 [2024-11-20 13:46:48.041495] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:41.104 [2024-11-20 13:46:48.076609] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:26:41.104 BaseBdev1 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:41.104 [ 00:26:41.104 { 00:26:41.104 "name": "BaseBdev1", 00:26:41.104 "aliases": [ 00:26:41.104 "a4e489a6-66d9-4883-b6d2-2d28e05bd73e" 00:26:41.104 ], 00:26:41.104 "product_name": "Malloc disk", 00:26:41.104 "block_size": 512, 00:26:41.104 "num_blocks": 65536, 00:26:41.104 "uuid": "a4e489a6-66d9-4883-b6d2-2d28e05bd73e", 00:26:41.104 "assigned_rate_limits": { 00:26:41.104 "rw_ios_per_sec": 0, 00:26:41.104 "rw_mbytes_per_sec": 0, 00:26:41.104 "r_mbytes_per_sec": 0, 00:26:41.104 "w_mbytes_per_sec": 0 00:26:41.104 }, 00:26:41.104 "claimed": true, 00:26:41.104 "claim_type": "exclusive_write", 00:26:41.104 "zoned": false, 00:26:41.104 "supported_io_types": { 00:26:41.104 "read": true, 00:26:41.104 "write": true, 00:26:41.104 "unmap": true, 00:26:41.104 "flush": true, 00:26:41.104 "reset": true, 00:26:41.104 "nvme_admin": false, 00:26:41.104 "nvme_io": false, 00:26:41.104 "nvme_io_md": false, 00:26:41.104 "write_zeroes": true, 00:26:41.104 "zcopy": true, 00:26:41.104 "get_zone_info": false, 00:26:41.104 "zone_management": false, 00:26:41.104 "zone_append": false, 00:26:41.104 "compare": false, 00:26:41.104 "compare_and_write": false, 00:26:41.104 "abort": true, 00:26:41.104 "seek_hole": false, 00:26:41.104 "seek_data": false, 00:26:41.104 "copy": true, 00:26:41.104 "nvme_iov_md": false 00:26:41.104 }, 00:26:41.104 "memory_domains": [ 00:26:41.104 { 00:26:41.104 "dma_device_id": "system", 00:26:41.104 "dma_device_type": 1 00:26:41.104 }, 00:26:41.104 { 00:26:41.104 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:41.104 "dma_device_type": 2 00:26:41.104 } 00:26:41.104 ], 00:26:41.104 "driver_specific": {} 00:26:41.104 } 00:26:41.104 ] 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:26:41.104 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:41.105 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:41.105 "name": "Existed_Raid", 00:26:41.105 "uuid": "bf667ac5-d676-4030-a479-ab0b09c201fc", 00:26:41.105 "strip_size_kb": 64, 00:26:41.105 "state": "configuring", 00:26:41.105 "raid_level": "raid0", 00:26:41.105 "superblock": true, 00:26:41.105 "num_base_bdevs": 2, 00:26:41.105 "num_base_bdevs_discovered": 1, 00:26:41.105 "num_base_bdevs_operational": 2, 00:26:41.105 "base_bdevs_list": [ 00:26:41.105 { 00:26:41.105 "name": "BaseBdev1", 00:26:41.105 "uuid": "a4e489a6-66d9-4883-b6d2-2d28e05bd73e", 00:26:41.105 "is_configured": true, 00:26:41.105 "data_offset": 2048, 00:26:41.105 "data_size": 63488 00:26:41.105 }, 00:26:41.105 { 00:26:41.105 "name": "BaseBdev2", 00:26:41.105 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:41.105 "is_configured": false, 00:26:41.105 "data_offset": 0, 00:26:41.105 "data_size": 0 00:26:41.105 } 00:26:41.105 ] 00:26:41.105 }' 00:26:41.105 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:41.105 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:41.363 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:26:41.363 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:41.363 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:41.363 [2024-11-20 13:46:48.408747] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:26:41.363 [2024-11-20 13:46:48.408809] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:26:41.363 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:41.363 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:26:41.363 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:41.363 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:41.363 [2024-11-20 13:46:48.416797] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:26:41.363 [2024-11-20 13:46:48.418828] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:26:41.363 [2024-11-20 13:46:48.418874] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:26:41.621 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:41.621 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:26:41.621 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:26:41.621 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:26:41.621 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:26:41.621 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:26:41.621 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:26:41.621 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:26:41.621 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:41.621 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:41.621 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:41.621 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:41.621 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:41.621 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:41.621 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:26:41.621 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:41.621 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:41.621 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:41.621 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:41.621 "name": "Existed_Raid", 00:26:41.621 "uuid": "b14eced8-4705-431b-ba87-628b0e4d6b23", 00:26:41.621 "strip_size_kb": 64, 00:26:41.621 "state": "configuring", 00:26:41.621 "raid_level": "raid0", 00:26:41.621 "superblock": true, 00:26:41.621 "num_base_bdevs": 2, 00:26:41.621 "num_base_bdevs_discovered": 1, 00:26:41.621 "num_base_bdevs_operational": 2, 00:26:41.621 "base_bdevs_list": [ 00:26:41.621 { 00:26:41.621 "name": "BaseBdev1", 00:26:41.621 "uuid": "a4e489a6-66d9-4883-b6d2-2d28e05bd73e", 00:26:41.621 "is_configured": true, 00:26:41.621 "data_offset": 2048, 00:26:41.621 "data_size": 63488 00:26:41.621 }, 00:26:41.621 { 00:26:41.621 "name": "BaseBdev2", 00:26:41.621 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:41.621 "is_configured": false, 00:26:41.621 "data_offset": 0, 00:26:41.621 "data_size": 0 00:26:41.621 } 00:26:41.621 ] 00:26:41.621 }' 00:26:41.621 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:41.621 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:41.879 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:26:41.879 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:41.879 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:41.879 [2024-11-20 13:46:48.749815] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:26:41.879 [2024-11-20 13:46:48.750095] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:26:41.879 [2024-11-20 13:46:48.750110] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:26:41.879 [2024-11-20 13:46:48.750384] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:26:41.879 BaseBdev2 00:26:41.879 [2024-11-20 13:46:48.750537] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:26:41.879 [2024-11-20 13:46:48.750549] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:26:41.879 [2024-11-20 13:46:48.750680] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:41.879 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:41.879 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:26:41.879 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:26:41.879 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:26:41.879 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:26:41.879 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:26:41.879 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:26:41.879 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:26:41.879 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:41.879 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:41.879 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:41.879 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:26:41.879 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:41.879 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:41.879 [ 00:26:41.879 { 00:26:41.879 "name": "BaseBdev2", 00:26:41.879 "aliases": [ 00:26:41.879 "079793fb-f508-461c-8f33-8f80bc0fcaa7" 00:26:41.879 ], 00:26:41.879 "product_name": "Malloc disk", 00:26:41.879 "block_size": 512, 00:26:41.879 "num_blocks": 65536, 00:26:41.879 "uuid": "079793fb-f508-461c-8f33-8f80bc0fcaa7", 00:26:41.879 "assigned_rate_limits": { 00:26:41.879 "rw_ios_per_sec": 0, 00:26:41.879 "rw_mbytes_per_sec": 0, 00:26:41.879 "r_mbytes_per_sec": 0, 00:26:41.880 "w_mbytes_per_sec": 0 00:26:41.880 }, 00:26:41.880 "claimed": true, 00:26:41.880 "claim_type": "exclusive_write", 00:26:41.880 "zoned": false, 00:26:41.880 "supported_io_types": { 00:26:41.880 "read": true, 00:26:41.880 "write": true, 00:26:41.880 "unmap": true, 00:26:41.880 "flush": true, 00:26:41.880 "reset": true, 00:26:41.880 "nvme_admin": false, 00:26:41.880 "nvme_io": false, 00:26:41.880 "nvme_io_md": false, 00:26:41.880 "write_zeroes": true, 00:26:41.880 "zcopy": true, 00:26:41.880 "get_zone_info": false, 00:26:41.880 "zone_management": false, 00:26:41.880 "zone_append": false, 00:26:41.880 "compare": false, 00:26:41.880 "compare_and_write": false, 00:26:41.880 "abort": true, 00:26:41.880 "seek_hole": false, 00:26:41.880 "seek_data": false, 00:26:41.880 "copy": true, 00:26:41.880 "nvme_iov_md": false 00:26:41.880 }, 00:26:41.880 "memory_domains": [ 00:26:41.880 { 00:26:41.880 "dma_device_id": "system", 00:26:41.880 "dma_device_type": 1 00:26:41.880 }, 00:26:41.880 { 00:26:41.880 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:41.880 "dma_device_type": 2 00:26:41.880 } 00:26:41.880 ], 00:26:41.880 "driver_specific": {} 00:26:41.880 } 00:26:41.880 ] 00:26:41.880 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:41.880 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:26:41.880 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:26:41.880 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:26:41.880 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 2 00:26:41.880 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:26:41.880 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:26:41.880 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:26:41.880 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:26:41.880 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:41.880 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:41.880 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:41.880 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:41.880 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:41.880 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:26:41.880 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:41.880 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:41.880 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:41.880 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:41.880 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:41.880 "name": "Existed_Raid", 00:26:41.880 "uuid": "b14eced8-4705-431b-ba87-628b0e4d6b23", 00:26:41.880 "strip_size_kb": 64, 00:26:41.880 "state": "online", 00:26:41.880 "raid_level": "raid0", 00:26:41.880 "superblock": true, 00:26:41.880 "num_base_bdevs": 2, 00:26:41.880 "num_base_bdevs_discovered": 2, 00:26:41.880 "num_base_bdevs_operational": 2, 00:26:41.880 "base_bdevs_list": [ 00:26:41.880 { 00:26:41.880 "name": "BaseBdev1", 00:26:41.880 "uuid": "a4e489a6-66d9-4883-b6d2-2d28e05bd73e", 00:26:41.880 "is_configured": true, 00:26:41.880 "data_offset": 2048, 00:26:41.880 "data_size": 63488 00:26:41.880 }, 00:26:41.880 { 00:26:41.880 "name": "BaseBdev2", 00:26:41.880 "uuid": "079793fb-f508-461c-8f33-8f80bc0fcaa7", 00:26:41.880 "is_configured": true, 00:26:41.880 "data_offset": 2048, 00:26:41.880 "data_size": 63488 00:26:41.880 } 00:26:41.880 ] 00:26:41.880 }' 00:26:41.880 13:46:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:41.880 13:46:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:42.139 [2024-11-20 13:46:49.082281] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:26:42.139 "name": "Existed_Raid", 00:26:42.139 "aliases": [ 00:26:42.139 "b14eced8-4705-431b-ba87-628b0e4d6b23" 00:26:42.139 ], 00:26:42.139 "product_name": "Raid Volume", 00:26:42.139 "block_size": 512, 00:26:42.139 "num_blocks": 126976, 00:26:42.139 "uuid": "b14eced8-4705-431b-ba87-628b0e4d6b23", 00:26:42.139 "assigned_rate_limits": { 00:26:42.139 "rw_ios_per_sec": 0, 00:26:42.139 "rw_mbytes_per_sec": 0, 00:26:42.139 "r_mbytes_per_sec": 0, 00:26:42.139 "w_mbytes_per_sec": 0 00:26:42.139 }, 00:26:42.139 "claimed": false, 00:26:42.139 "zoned": false, 00:26:42.139 "supported_io_types": { 00:26:42.139 "read": true, 00:26:42.139 "write": true, 00:26:42.139 "unmap": true, 00:26:42.139 "flush": true, 00:26:42.139 "reset": true, 00:26:42.139 "nvme_admin": false, 00:26:42.139 "nvme_io": false, 00:26:42.139 "nvme_io_md": false, 00:26:42.139 "write_zeroes": true, 00:26:42.139 "zcopy": false, 00:26:42.139 "get_zone_info": false, 00:26:42.139 "zone_management": false, 00:26:42.139 "zone_append": false, 00:26:42.139 "compare": false, 00:26:42.139 "compare_and_write": false, 00:26:42.139 "abort": false, 00:26:42.139 "seek_hole": false, 00:26:42.139 "seek_data": false, 00:26:42.139 "copy": false, 00:26:42.139 "nvme_iov_md": false 00:26:42.139 }, 00:26:42.139 "memory_domains": [ 00:26:42.139 { 00:26:42.139 "dma_device_id": "system", 00:26:42.139 "dma_device_type": 1 00:26:42.139 }, 00:26:42.139 { 00:26:42.139 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:42.139 "dma_device_type": 2 00:26:42.139 }, 00:26:42.139 { 00:26:42.139 "dma_device_id": "system", 00:26:42.139 "dma_device_type": 1 00:26:42.139 }, 00:26:42.139 { 00:26:42.139 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:42.139 "dma_device_type": 2 00:26:42.139 } 00:26:42.139 ], 00:26:42.139 "driver_specific": { 00:26:42.139 "raid": { 00:26:42.139 "uuid": "b14eced8-4705-431b-ba87-628b0e4d6b23", 00:26:42.139 "strip_size_kb": 64, 00:26:42.139 "state": "online", 00:26:42.139 "raid_level": "raid0", 00:26:42.139 "superblock": true, 00:26:42.139 "num_base_bdevs": 2, 00:26:42.139 "num_base_bdevs_discovered": 2, 00:26:42.139 "num_base_bdevs_operational": 2, 00:26:42.139 "base_bdevs_list": [ 00:26:42.139 { 00:26:42.139 "name": "BaseBdev1", 00:26:42.139 "uuid": "a4e489a6-66d9-4883-b6d2-2d28e05bd73e", 00:26:42.139 "is_configured": true, 00:26:42.139 "data_offset": 2048, 00:26:42.139 "data_size": 63488 00:26:42.139 }, 00:26:42.139 { 00:26:42.139 "name": "BaseBdev2", 00:26:42.139 "uuid": "079793fb-f508-461c-8f33-8f80bc0fcaa7", 00:26:42.139 "is_configured": true, 00:26:42.139 "data_offset": 2048, 00:26:42.139 "data_size": 63488 00:26:42.139 } 00:26:42.139 ] 00:26:42.139 } 00:26:42.139 } 00:26:42.139 }' 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:26:42.139 BaseBdev2' 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:42.139 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:42.397 [2024-11-20 13:46:49.222009] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:26:42.397 [2024-11-20 13:46:49.222047] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:26:42.397 [2024-11-20 13:46:49.222104] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 1 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:42.397 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:42.397 "name": "Existed_Raid", 00:26:42.397 "uuid": "b14eced8-4705-431b-ba87-628b0e4d6b23", 00:26:42.397 "strip_size_kb": 64, 00:26:42.397 "state": "offline", 00:26:42.397 "raid_level": "raid0", 00:26:42.397 "superblock": true, 00:26:42.397 "num_base_bdevs": 2, 00:26:42.397 "num_base_bdevs_discovered": 1, 00:26:42.397 "num_base_bdevs_operational": 1, 00:26:42.397 "base_bdevs_list": [ 00:26:42.397 { 00:26:42.398 "name": null, 00:26:42.398 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:42.398 "is_configured": false, 00:26:42.398 "data_offset": 0, 00:26:42.398 "data_size": 63488 00:26:42.398 }, 00:26:42.398 { 00:26:42.398 "name": "BaseBdev2", 00:26:42.398 "uuid": "079793fb-f508-461c-8f33-8f80bc0fcaa7", 00:26:42.398 "is_configured": true, 00:26:42.398 "data_offset": 2048, 00:26:42.398 "data_size": 63488 00:26:42.398 } 00:26:42.398 ] 00:26:42.398 }' 00:26:42.398 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:42.398 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:42.656 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:26:42.656 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:26:42.656 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:42.656 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:26:42.656 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:42.656 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:42.656 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:42.656 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:26:42.656 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:26:42.656 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:26:42.656 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:42.656 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:42.656 [2024-11-20 13:46:49.617011] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:26:42.656 [2024-11-20 13:46:49.617064] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:26:42.656 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:42.656 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:26:42.656 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:26:42.656 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:42.656 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:26:42.656 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:42.656 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:42.656 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:42.915 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:26:42.915 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:26:42.915 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:26:42.915 13:46:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 59643 00:26:42.915 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 59643 ']' 00:26:42.915 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 59643 00:26:42.915 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:26:42.915 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:42.915 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 59643 00:26:42.915 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:26:42.915 killing process with pid 59643 00:26:42.915 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:26:42.915 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 59643' 00:26:42.915 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 59643 00:26:42.915 [2024-11-20 13:46:49.741532] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:26:42.915 13:46:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 59643 00:26:42.915 [2024-11-20 13:46:49.752619] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:26:43.481 13:46:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:26:43.481 00:26:43.481 real 0m3.744s 00:26:43.481 user 0m5.318s 00:26:43.481 sys 0m0.602s 00:26:43.481 13:46:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:26:43.481 ************************************ 00:26:43.481 END TEST raid_state_function_test_sb 00:26:43.481 ************************************ 00:26:43.481 13:46:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:43.740 13:46:50 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 2 00:26:43.740 13:46:50 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:26:43.740 13:46:50 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:26:43.740 13:46:50 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:26:43.740 ************************************ 00:26:43.740 START TEST raid_superblock_test 00:26:43.740 ************************************ 00:26:43.740 13:46:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid0 2 00:26:43.740 13:46:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:26:43.740 13:46:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:26:43.740 13:46:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:26:43.740 13:46:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:26:43.740 13:46:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:26:43.740 13:46:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:26:43.740 13:46:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:26:43.740 13:46:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:26:43.740 13:46:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:26:43.740 13:46:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:26:43.740 13:46:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:26:43.740 13:46:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:26:43.740 13:46:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:26:43.740 13:46:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:26:43.740 13:46:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:26:43.740 13:46:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:26:43.740 13:46:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=59879 00:26:43.740 13:46:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 59879 00:26:43.740 13:46:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 59879 ']' 00:26:43.740 13:46:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:43.740 13:46:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:43.740 13:46:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:43.740 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:43.740 13:46:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:26:43.740 13:46:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:43.740 13:46:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:43.740 [2024-11-20 13:46:50.639814] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:26:43.740 [2024-11-20 13:46:50.639967] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid59879 ] 00:26:43.999 [2024-11-20 13:46:50.803787] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:43.999 [2024-11-20 13:46:50.993151] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:44.258 [2024-11-20 13:46:51.164121] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:44.258 [2024-11-20 13:46:51.164183] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:44.516 malloc1 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:44.516 [2024-11-20 13:46:51.534212] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:26:44.516 [2024-11-20 13:46:51.534276] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:44.516 [2024-11-20 13:46:51.534300] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:26:44.516 [2024-11-20 13:46:51.534310] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:44.516 [2024-11-20 13:46:51.536627] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:44.516 [2024-11-20 13:46:51.536665] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:26:44.516 pt1 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:44.516 malloc2 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:44.516 13:46:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:44.775 [2024-11-20 13:46:51.572970] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:26:44.775 [2024-11-20 13:46:51.573050] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:44.775 [2024-11-20 13:46:51.573077] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:26:44.775 [2024-11-20 13:46:51.573086] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:44.775 [2024-11-20 13:46:51.575408] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:44.775 [2024-11-20 13:46:51.575443] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:26:44.775 pt2 00:26:44.775 13:46:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:44.775 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:26:44.775 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:26:44.775 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:26:44.775 13:46:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:44.775 13:46:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:44.775 [2024-11-20 13:46:51.581044] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:26:44.775 [2024-11-20 13:46:51.583060] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:26:44.775 [2024-11-20 13:46:51.583225] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:26:44.775 [2024-11-20 13:46:51.583237] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:26:44.775 [2024-11-20 13:46:51.583549] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:26:44.775 [2024-11-20 13:46:51.583697] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:26:44.775 [2024-11-20 13:46:51.583709] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:26:44.775 [2024-11-20 13:46:51.583868] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:44.775 13:46:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:44.775 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:26:44.775 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:26:44.775 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:26:44.775 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:26:44.775 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:26:44.775 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:44.775 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:44.775 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:44.775 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:44.775 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:44.775 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:44.775 13:46:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:44.775 13:46:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:44.775 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:44.775 13:46:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:44.775 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:44.775 "name": "raid_bdev1", 00:26:44.775 "uuid": "3aa22a26-8b8a-4a04-8447-0293079ba58f", 00:26:44.775 "strip_size_kb": 64, 00:26:44.775 "state": "online", 00:26:44.775 "raid_level": "raid0", 00:26:44.775 "superblock": true, 00:26:44.775 "num_base_bdevs": 2, 00:26:44.775 "num_base_bdevs_discovered": 2, 00:26:44.775 "num_base_bdevs_operational": 2, 00:26:44.775 "base_bdevs_list": [ 00:26:44.775 { 00:26:44.775 "name": "pt1", 00:26:44.775 "uuid": "00000000-0000-0000-0000-000000000001", 00:26:44.775 "is_configured": true, 00:26:44.775 "data_offset": 2048, 00:26:44.775 "data_size": 63488 00:26:44.775 }, 00:26:44.775 { 00:26:44.775 "name": "pt2", 00:26:44.775 "uuid": "00000000-0000-0000-0000-000000000002", 00:26:44.775 "is_configured": true, 00:26:44.775 "data_offset": 2048, 00:26:44.775 "data_size": 63488 00:26:44.775 } 00:26:44.775 ] 00:26:44.775 }' 00:26:44.775 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:44.775 13:46:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:45.033 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:26:45.033 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:26:45.034 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:26:45.034 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:26:45.034 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:26:45.034 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:26:45.034 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:26:45.034 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:26:45.034 13:46:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:45.034 13:46:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:45.034 [2024-11-20 13:46:51.933397] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:45.034 13:46:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:45.034 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:26:45.034 "name": "raid_bdev1", 00:26:45.034 "aliases": [ 00:26:45.034 "3aa22a26-8b8a-4a04-8447-0293079ba58f" 00:26:45.034 ], 00:26:45.034 "product_name": "Raid Volume", 00:26:45.034 "block_size": 512, 00:26:45.034 "num_blocks": 126976, 00:26:45.034 "uuid": "3aa22a26-8b8a-4a04-8447-0293079ba58f", 00:26:45.034 "assigned_rate_limits": { 00:26:45.034 "rw_ios_per_sec": 0, 00:26:45.034 "rw_mbytes_per_sec": 0, 00:26:45.034 "r_mbytes_per_sec": 0, 00:26:45.034 "w_mbytes_per_sec": 0 00:26:45.034 }, 00:26:45.034 "claimed": false, 00:26:45.034 "zoned": false, 00:26:45.034 "supported_io_types": { 00:26:45.034 "read": true, 00:26:45.034 "write": true, 00:26:45.034 "unmap": true, 00:26:45.034 "flush": true, 00:26:45.034 "reset": true, 00:26:45.034 "nvme_admin": false, 00:26:45.034 "nvme_io": false, 00:26:45.034 "nvme_io_md": false, 00:26:45.034 "write_zeroes": true, 00:26:45.034 "zcopy": false, 00:26:45.034 "get_zone_info": false, 00:26:45.034 "zone_management": false, 00:26:45.034 "zone_append": false, 00:26:45.034 "compare": false, 00:26:45.034 "compare_and_write": false, 00:26:45.034 "abort": false, 00:26:45.034 "seek_hole": false, 00:26:45.034 "seek_data": false, 00:26:45.034 "copy": false, 00:26:45.034 "nvme_iov_md": false 00:26:45.034 }, 00:26:45.034 "memory_domains": [ 00:26:45.034 { 00:26:45.034 "dma_device_id": "system", 00:26:45.034 "dma_device_type": 1 00:26:45.034 }, 00:26:45.034 { 00:26:45.034 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:45.034 "dma_device_type": 2 00:26:45.034 }, 00:26:45.034 { 00:26:45.034 "dma_device_id": "system", 00:26:45.034 "dma_device_type": 1 00:26:45.034 }, 00:26:45.034 { 00:26:45.034 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:45.034 "dma_device_type": 2 00:26:45.034 } 00:26:45.034 ], 00:26:45.034 "driver_specific": { 00:26:45.034 "raid": { 00:26:45.034 "uuid": "3aa22a26-8b8a-4a04-8447-0293079ba58f", 00:26:45.034 "strip_size_kb": 64, 00:26:45.034 "state": "online", 00:26:45.034 "raid_level": "raid0", 00:26:45.034 "superblock": true, 00:26:45.034 "num_base_bdevs": 2, 00:26:45.034 "num_base_bdevs_discovered": 2, 00:26:45.034 "num_base_bdevs_operational": 2, 00:26:45.034 "base_bdevs_list": [ 00:26:45.034 { 00:26:45.034 "name": "pt1", 00:26:45.034 "uuid": "00000000-0000-0000-0000-000000000001", 00:26:45.034 "is_configured": true, 00:26:45.034 "data_offset": 2048, 00:26:45.034 "data_size": 63488 00:26:45.034 }, 00:26:45.034 { 00:26:45.034 "name": "pt2", 00:26:45.034 "uuid": "00000000-0000-0000-0000-000000000002", 00:26:45.034 "is_configured": true, 00:26:45.034 "data_offset": 2048, 00:26:45.034 "data_size": 63488 00:26:45.034 } 00:26:45.034 ] 00:26:45.034 } 00:26:45.034 } 00:26:45.034 }' 00:26:45.034 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:26:45.034 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:26:45.034 pt2' 00:26:45.034 13:46:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:26:45.034 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:26:45.034 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:26:45.034 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:26:45.034 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:45.034 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:45.034 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:26:45.034 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:45.034 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:26:45.034 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:26:45.034 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:26:45.034 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:26:45.034 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:26:45.034 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:45.034 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:45.034 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:45.292 [2024-11-20 13:46:52.105422] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=3aa22a26-8b8a-4a04-8447-0293079ba58f 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 3aa22a26-8b8a-4a04-8447-0293079ba58f ']' 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:45.292 [2024-11-20 13:46:52.137096] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:26:45.292 [2024-11-20 13:46:52.137121] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:26:45.292 [2024-11-20 13:46:52.137211] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:45.292 [2024-11-20 13:46:52.137266] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:26:45.292 [2024-11-20 13:46:52.137279] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:45.292 [2024-11-20 13:46:52.237164] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:26:45.292 [2024-11-20 13:46:52.239213] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:26:45.292 [2024-11-20 13:46:52.239283] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:26:45.292 [2024-11-20 13:46:52.239338] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:26:45.292 [2024-11-20 13:46:52.239362] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:26:45.292 [2024-11-20 13:46:52.239377] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:26:45.292 request: 00:26:45.292 { 00:26:45.292 "name": "raid_bdev1", 00:26:45.292 "raid_level": "raid0", 00:26:45.292 "base_bdevs": [ 00:26:45.292 "malloc1", 00:26:45.292 "malloc2" 00:26:45.292 ], 00:26:45.292 "strip_size_kb": 64, 00:26:45.292 "superblock": false, 00:26:45.292 "method": "bdev_raid_create", 00:26:45.292 "req_id": 1 00:26:45.292 } 00:26:45.292 Got JSON-RPC error response 00:26:45.292 response: 00:26:45.292 { 00:26:45.292 "code": -17, 00:26:45.292 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:26:45.292 } 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:45.292 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:45.293 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:26:45.293 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:45.293 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:26:45.293 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:26:45.293 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:26:45.293 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:45.293 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:45.293 [2024-11-20 13:46:52.285158] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:26:45.293 [2024-11-20 13:46:52.285231] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:45.293 [2024-11-20 13:46:52.285250] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:26:45.293 [2024-11-20 13:46:52.285261] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:45.293 [2024-11-20 13:46:52.287631] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:45.293 [2024-11-20 13:46:52.287788] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:26:45.293 [2024-11-20 13:46:52.287888] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:26:45.293 [2024-11-20 13:46:52.287946] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:26:45.293 pt1 00:26:45.293 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:45.293 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 2 00:26:45.293 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:26:45.293 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:26:45.293 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:26:45.293 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:26:45.293 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:45.293 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:45.293 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:45.293 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:45.293 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:45.293 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:45.293 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:45.293 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:45.293 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:45.293 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:45.293 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:45.293 "name": "raid_bdev1", 00:26:45.293 "uuid": "3aa22a26-8b8a-4a04-8447-0293079ba58f", 00:26:45.293 "strip_size_kb": 64, 00:26:45.293 "state": "configuring", 00:26:45.293 "raid_level": "raid0", 00:26:45.293 "superblock": true, 00:26:45.293 "num_base_bdevs": 2, 00:26:45.293 "num_base_bdevs_discovered": 1, 00:26:45.293 "num_base_bdevs_operational": 2, 00:26:45.293 "base_bdevs_list": [ 00:26:45.293 { 00:26:45.293 "name": "pt1", 00:26:45.293 "uuid": "00000000-0000-0000-0000-000000000001", 00:26:45.293 "is_configured": true, 00:26:45.293 "data_offset": 2048, 00:26:45.293 "data_size": 63488 00:26:45.293 }, 00:26:45.293 { 00:26:45.293 "name": null, 00:26:45.293 "uuid": "00000000-0000-0000-0000-000000000002", 00:26:45.293 "is_configured": false, 00:26:45.293 "data_offset": 2048, 00:26:45.293 "data_size": 63488 00:26:45.293 } 00:26:45.293 ] 00:26:45.293 }' 00:26:45.293 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:45.293 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:45.858 [2024-11-20 13:46:52.625271] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:26:45.858 [2024-11-20 13:46:52.625360] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:45.858 [2024-11-20 13:46:52.625384] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:26:45.858 [2024-11-20 13:46:52.625396] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:45.858 [2024-11-20 13:46:52.625895] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:45.858 [2024-11-20 13:46:52.625921] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:26:45.858 [2024-11-20 13:46:52.626023] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:26:45.858 [2024-11-20 13:46:52.626053] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:26:45.858 [2024-11-20 13:46:52.626168] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:26:45.858 [2024-11-20 13:46:52.626180] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:26:45.858 [2024-11-20 13:46:52.626435] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:26:45.858 [2024-11-20 13:46:52.626564] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:26:45.858 [2024-11-20 13:46:52.626572] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:26:45.858 [2024-11-20 13:46:52.626706] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:45.858 pt2 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:45.858 "name": "raid_bdev1", 00:26:45.858 "uuid": "3aa22a26-8b8a-4a04-8447-0293079ba58f", 00:26:45.858 "strip_size_kb": 64, 00:26:45.858 "state": "online", 00:26:45.858 "raid_level": "raid0", 00:26:45.858 "superblock": true, 00:26:45.858 "num_base_bdevs": 2, 00:26:45.858 "num_base_bdevs_discovered": 2, 00:26:45.858 "num_base_bdevs_operational": 2, 00:26:45.858 "base_bdevs_list": [ 00:26:45.858 { 00:26:45.858 "name": "pt1", 00:26:45.858 "uuid": "00000000-0000-0000-0000-000000000001", 00:26:45.858 "is_configured": true, 00:26:45.858 "data_offset": 2048, 00:26:45.858 "data_size": 63488 00:26:45.858 }, 00:26:45.858 { 00:26:45.858 "name": "pt2", 00:26:45.858 "uuid": "00000000-0000-0000-0000-000000000002", 00:26:45.858 "is_configured": true, 00:26:45.858 "data_offset": 2048, 00:26:45.858 "data_size": 63488 00:26:45.858 } 00:26:45.858 ] 00:26:45.858 }' 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:45.858 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:46.116 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:26:46.116 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:26:46.116 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:26:46.116 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:26:46.116 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:26:46.116 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:26:46.116 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:26:46.116 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:46.116 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:46.116 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:26:46.116 [2024-11-20 13:46:52.945599] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:46.116 13:46:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:46.116 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:26:46.116 "name": "raid_bdev1", 00:26:46.116 "aliases": [ 00:26:46.116 "3aa22a26-8b8a-4a04-8447-0293079ba58f" 00:26:46.116 ], 00:26:46.116 "product_name": "Raid Volume", 00:26:46.116 "block_size": 512, 00:26:46.116 "num_blocks": 126976, 00:26:46.116 "uuid": "3aa22a26-8b8a-4a04-8447-0293079ba58f", 00:26:46.116 "assigned_rate_limits": { 00:26:46.116 "rw_ios_per_sec": 0, 00:26:46.116 "rw_mbytes_per_sec": 0, 00:26:46.116 "r_mbytes_per_sec": 0, 00:26:46.116 "w_mbytes_per_sec": 0 00:26:46.117 }, 00:26:46.117 "claimed": false, 00:26:46.117 "zoned": false, 00:26:46.117 "supported_io_types": { 00:26:46.117 "read": true, 00:26:46.117 "write": true, 00:26:46.117 "unmap": true, 00:26:46.117 "flush": true, 00:26:46.117 "reset": true, 00:26:46.117 "nvme_admin": false, 00:26:46.117 "nvme_io": false, 00:26:46.117 "nvme_io_md": false, 00:26:46.117 "write_zeroes": true, 00:26:46.117 "zcopy": false, 00:26:46.117 "get_zone_info": false, 00:26:46.117 "zone_management": false, 00:26:46.117 "zone_append": false, 00:26:46.117 "compare": false, 00:26:46.117 "compare_and_write": false, 00:26:46.117 "abort": false, 00:26:46.117 "seek_hole": false, 00:26:46.117 "seek_data": false, 00:26:46.117 "copy": false, 00:26:46.117 "nvme_iov_md": false 00:26:46.117 }, 00:26:46.117 "memory_domains": [ 00:26:46.117 { 00:26:46.117 "dma_device_id": "system", 00:26:46.117 "dma_device_type": 1 00:26:46.117 }, 00:26:46.117 { 00:26:46.117 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:46.117 "dma_device_type": 2 00:26:46.117 }, 00:26:46.117 { 00:26:46.117 "dma_device_id": "system", 00:26:46.117 "dma_device_type": 1 00:26:46.117 }, 00:26:46.117 { 00:26:46.117 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:46.117 "dma_device_type": 2 00:26:46.117 } 00:26:46.117 ], 00:26:46.117 "driver_specific": { 00:26:46.117 "raid": { 00:26:46.117 "uuid": "3aa22a26-8b8a-4a04-8447-0293079ba58f", 00:26:46.117 "strip_size_kb": 64, 00:26:46.117 "state": "online", 00:26:46.117 "raid_level": "raid0", 00:26:46.117 "superblock": true, 00:26:46.117 "num_base_bdevs": 2, 00:26:46.117 "num_base_bdevs_discovered": 2, 00:26:46.117 "num_base_bdevs_operational": 2, 00:26:46.117 "base_bdevs_list": [ 00:26:46.117 { 00:26:46.117 "name": "pt1", 00:26:46.117 "uuid": "00000000-0000-0000-0000-000000000001", 00:26:46.117 "is_configured": true, 00:26:46.117 "data_offset": 2048, 00:26:46.117 "data_size": 63488 00:26:46.117 }, 00:26:46.117 { 00:26:46.117 "name": "pt2", 00:26:46.117 "uuid": "00000000-0000-0000-0000-000000000002", 00:26:46.117 "is_configured": true, 00:26:46.117 "data_offset": 2048, 00:26:46.117 "data_size": 63488 00:26:46.117 } 00:26:46.117 ] 00:26:46.117 } 00:26:46.117 } 00:26:46.117 }' 00:26:46.117 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:26:46.117 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:26:46.117 pt2' 00:26:46.117 13:46:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:46.117 [2024-11-20 13:46:53.097593] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 3aa22a26-8b8a-4a04-8447-0293079ba58f '!=' 3aa22a26-8b8a-4a04-8447-0293079ba58f ']' 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 59879 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 59879 ']' 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 59879 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 59879 00:26:46.117 killing process with pid 59879 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 59879' 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 59879 00:26:46.117 [2024-11-20 13:46:53.154830] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:26:46.117 [2024-11-20 13:46:53.154923] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:46.117 13:46:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 59879 00:26:46.117 [2024-11-20 13:46:53.155000] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:26:46.117 [2024-11-20 13:46:53.155017] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:26:46.375 [2024-11-20 13:46:53.290036] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:26:47.336 13:46:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:26:47.336 00:26:47.336 real 0m3.476s 00:26:47.336 user 0m4.831s 00:26:47.336 sys 0m0.575s 00:26:47.336 13:46:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:26:47.336 13:46:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:26:47.336 ************************************ 00:26:47.336 END TEST raid_superblock_test 00:26:47.336 ************************************ 00:26:47.336 13:46:54 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 2 read 00:26:47.336 13:46:54 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:26:47.336 13:46:54 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:26:47.336 13:46:54 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:26:47.336 ************************************ 00:26:47.336 START TEST raid_read_error_test 00:26:47.336 ************************************ 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 2 read 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:26:47.336 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.EK3pNCZmE6 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=60085 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 60085 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 60085 ']' 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:47.336 13:46:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:26:47.337 [2024-11-20 13:46:54.161863] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:26:47.337 [2024-11-20 13:46:54.162015] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid60085 ] 00:26:47.337 [2024-11-20 13:46:54.325575] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:47.595 [2024-11-20 13:46:54.448717] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:47.595 [2024-11-20 13:46:54.598461] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:47.595 [2024-11-20 13:46:54.598529] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:48.164 BaseBdev1_malloc 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:48.164 true 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:48.164 [2024-11-20 13:46:55.059793] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:26:48.164 [2024-11-20 13:46:55.059861] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:48.164 [2024-11-20 13:46:55.059883] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:26:48.164 [2024-11-20 13:46:55.059895] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:48.164 [2024-11-20 13:46:55.062164] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:48.164 [2024-11-20 13:46:55.062312] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:26:48.164 BaseBdev1 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:48.164 BaseBdev2_malloc 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:48.164 true 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:48.164 [2024-11-20 13:46:55.106246] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:26:48.164 [2024-11-20 13:46:55.106298] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:48.164 [2024-11-20 13:46:55.106314] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:26:48.164 [2024-11-20 13:46:55.106325] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:48.164 [2024-11-20 13:46:55.108543] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:48.164 [2024-11-20 13:46:55.108580] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:26:48.164 BaseBdev2 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:48.164 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:48.164 [2024-11-20 13:46:55.114308] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:26:48.165 [2024-11-20 13:46:55.116290] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:26:48.165 [2024-11-20 13:46:55.116475] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:26:48.165 [2024-11-20 13:46:55.116491] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:26:48.165 [2024-11-20 13:46:55.116739] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:26:48.165 [2024-11-20 13:46:55.116890] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:26:48.165 [2024-11-20 13:46:55.116901] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:26:48.165 [2024-11-20 13:46:55.117073] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:48.165 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:48.165 13:46:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:26:48.165 13:46:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:26:48.165 13:46:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:26:48.165 13:46:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:26:48.165 13:46:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:26:48.165 13:46:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:48.165 13:46:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:48.165 13:46:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:48.165 13:46:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:48.165 13:46:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:48.165 13:46:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:48.165 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:48.165 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:48.165 13:46:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:48.165 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:48.165 13:46:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:48.165 "name": "raid_bdev1", 00:26:48.165 "uuid": "9fea6bc7-63b1-40fc-b271-ee2e30392a98", 00:26:48.165 "strip_size_kb": 64, 00:26:48.165 "state": "online", 00:26:48.165 "raid_level": "raid0", 00:26:48.165 "superblock": true, 00:26:48.165 "num_base_bdevs": 2, 00:26:48.165 "num_base_bdevs_discovered": 2, 00:26:48.165 "num_base_bdevs_operational": 2, 00:26:48.165 "base_bdevs_list": [ 00:26:48.165 { 00:26:48.165 "name": "BaseBdev1", 00:26:48.165 "uuid": "eb903c6d-3893-53ec-aa92-ddddf0284f46", 00:26:48.165 "is_configured": true, 00:26:48.165 "data_offset": 2048, 00:26:48.165 "data_size": 63488 00:26:48.165 }, 00:26:48.165 { 00:26:48.165 "name": "BaseBdev2", 00:26:48.165 "uuid": "ea4100e6-9f18-50c6-ba0e-00f52732b293", 00:26:48.165 "is_configured": true, 00:26:48.165 "data_offset": 2048, 00:26:48.165 "data_size": 63488 00:26:48.165 } 00:26:48.165 ] 00:26:48.165 }' 00:26:48.165 13:46:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:48.165 13:46:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:48.423 13:46:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:26:48.423 13:46:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:26:48.681 [2024-11-20 13:46:55.531480] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:26:49.618 13:46:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:26:49.618 13:46:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:49.618 13:46:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:49.618 13:46:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:49.618 13:46:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:26:49.618 13:46:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:26:49.618 13:46:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:26:49.618 13:46:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:26:49.618 13:46:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:26:49.618 13:46:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:26:49.618 13:46:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:26:49.618 13:46:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:26:49.618 13:46:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:49.618 13:46:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:49.618 13:46:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:49.618 13:46:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:49.618 13:46:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:49.618 13:46:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:49.618 13:46:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:49.618 13:46:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:49.619 13:46:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:49.619 13:46:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:49.619 13:46:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:49.619 "name": "raid_bdev1", 00:26:49.619 "uuid": "9fea6bc7-63b1-40fc-b271-ee2e30392a98", 00:26:49.619 "strip_size_kb": 64, 00:26:49.619 "state": "online", 00:26:49.619 "raid_level": "raid0", 00:26:49.619 "superblock": true, 00:26:49.619 "num_base_bdevs": 2, 00:26:49.619 "num_base_bdevs_discovered": 2, 00:26:49.619 "num_base_bdevs_operational": 2, 00:26:49.619 "base_bdevs_list": [ 00:26:49.619 { 00:26:49.619 "name": "BaseBdev1", 00:26:49.619 "uuid": "eb903c6d-3893-53ec-aa92-ddddf0284f46", 00:26:49.619 "is_configured": true, 00:26:49.619 "data_offset": 2048, 00:26:49.619 "data_size": 63488 00:26:49.619 }, 00:26:49.619 { 00:26:49.619 "name": "BaseBdev2", 00:26:49.619 "uuid": "ea4100e6-9f18-50c6-ba0e-00f52732b293", 00:26:49.619 "is_configured": true, 00:26:49.619 "data_offset": 2048, 00:26:49.619 "data_size": 63488 00:26:49.619 } 00:26:49.619 ] 00:26:49.619 }' 00:26:49.619 13:46:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:49.619 13:46:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:49.879 13:46:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:26:49.879 13:46:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:49.879 13:46:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:49.879 [2024-11-20 13:46:56.805974] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:26:49.879 [2024-11-20 13:46:56.806025] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:26:49.879 [2024-11-20 13:46:56.809223] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:49.879 [2024-11-20 13:46:56.809274] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:49.879 [2024-11-20 13:46:56.809310] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:26:49.879 [2024-11-20 13:46:56.809322] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:26:49.879 { 00:26:49.879 "results": [ 00:26:49.879 { 00:26:49.879 "job": "raid_bdev1", 00:26:49.879 "core_mask": "0x1", 00:26:49.879 "workload": "randrw", 00:26:49.879 "percentage": 50, 00:26:49.879 "status": "finished", 00:26:49.879 "queue_depth": 1, 00:26:49.879 "io_size": 131072, 00:26:49.879 "runtime": 1.272583, 00:26:49.879 "iops": 13848.998454324787, 00:26:49.879 "mibps": 1731.1248067905983, 00:26:49.879 "io_failed": 1, 00:26:49.879 "io_timeout": 0, 00:26:49.879 "avg_latency_us": 98.87903773049645, 00:26:49.879 "min_latency_us": 34.067692307692305, 00:26:49.879 "max_latency_us": 1701.4153846153847 00:26:49.879 } 00:26:49.879 ], 00:26:49.879 "core_count": 1 00:26:49.879 } 00:26:49.879 13:46:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:49.879 13:46:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 60085 00:26:49.879 13:46:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 60085 ']' 00:26:49.879 13:46:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 60085 00:26:49.879 13:46:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:26:49.879 13:46:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:49.879 13:46:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 60085 00:26:49.879 killing process with pid 60085 00:26:49.879 13:46:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:26:49.879 13:46:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:26:49.879 13:46:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 60085' 00:26:49.879 13:46:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 60085 00:26:49.879 13:46:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 60085 00:26:49.879 [2024-11-20 13:46:56.839848] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:26:49.879 [2024-11-20 13:46:56.929799] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:26:50.816 13:46:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:26:50.816 13:46:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.EK3pNCZmE6 00:26:50.816 13:46:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:26:50.816 13:46:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.79 00:26:50.816 13:46:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:26:50.816 13:46:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:26:50.816 13:46:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:26:50.816 ************************************ 00:26:50.816 END TEST raid_read_error_test 00:26:50.816 ************************************ 00:26:50.816 13:46:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.79 != \0\.\0\0 ]] 00:26:50.816 00:26:50.816 real 0m3.666s 00:26:50.816 user 0m4.337s 00:26:50.816 sys 0m0.439s 00:26:50.816 13:46:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:26:50.816 13:46:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:50.816 13:46:57 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 2 write 00:26:50.816 13:46:57 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:26:50.816 13:46:57 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:26:50.816 13:46:57 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:26:50.816 ************************************ 00:26:50.816 START TEST raid_write_error_test 00:26:50.816 ************************************ 00:26:50.816 13:46:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 2 write 00:26:50.816 13:46:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:26:50.816 13:46:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:26:50.816 13:46:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:26:50.816 13:46:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:26:50.816 13:46:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:26:50.816 13:46:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:26:50.816 13:46:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:26:50.816 13:46:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:26:50.816 13:46:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:26:50.816 13:46:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:26:50.817 13:46:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:26:50.817 13:46:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:26:50.817 13:46:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:26:50.817 13:46:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:26:50.817 13:46:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:26:50.817 13:46:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:26:50.817 13:46:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:26:50.817 13:46:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:26:50.817 13:46:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:26:50.817 13:46:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:26:50.817 13:46:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:26:50.817 13:46:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:26:50.817 13:46:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.4txNlrfn0o 00:26:50.817 13:46:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=60216 00:26:50.817 13:46:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 60216 00:26:50.817 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:50.817 13:46:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 60216 ']' 00:26:50.817 13:46:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:50.817 13:46:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:26:50.817 13:46:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:50.817 13:46:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:50.817 13:46:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:50.817 13:46:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:50.817 [2024-11-20 13:46:57.866681] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:26:50.817 [2024-11-20 13:46:57.866820] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid60216 ] 00:26:51.077 [2024-11-20 13:46:58.028058] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:51.337 [2024-11-20 13:46:58.148186] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:51.337 [2024-11-20 13:46:58.296125] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:51.337 [2024-11-20 13:46:58.296169] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:51.905 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:51.905 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:26:51.905 13:46:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:26:51.905 13:46:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:26:51.905 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:51.906 BaseBdev1_malloc 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:51.906 true 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:51.906 [2024-11-20 13:46:58.755489] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:26:51.906 [2024-11-20 13:46:58.755676] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:51.906 [2024-11-20 13:46:58.755704] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:26:51.906 [2024-11-20 13:46:58.755716] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:51.906 [2024-11-20 13:46:58.757963] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:51.906 [2024-11-20 13:46:58.758012] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:26:51.906 BaseBdev1 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:51.906 BaseBdev2_malloc 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:51.906 true 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:51.906 [2024-11-20 13:46:58.801622] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:26:51.906 [2024-11-20 13:46:58.801686] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:51.906 [2024-11-20 13:46:58.801704] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:26:51.906 [2024-11-20 13:46:58.801715] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:51.906 [2024-11-20 13:46:58.804055] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:51.906 [2024-11-20 13:46:58.804092] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:26:51.906 BaseBdev2 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:51.906 [2024-11-20 13:46:58.809683] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:26:51.906 [2024-11-20 13:46:58.811646] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:26:51.906 [2024-11-20 13:46:58.811830] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:26:51.906 [2024-11-20 13:46:58.811846] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:26:51.906 [2024-11-20 13:46:58.812134] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:26:51.906 [2024-11-20 13:46:58.812296] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:26:51.906 [2024-11-20 13:46:58.812307] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:26:51.906 [2024-11-20 13:46:58.812450] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:51.906 "name": "raid_bdev1", 00:26:51.906 "uuid": "1d5535f9-4a26-4f81-8c6e-5f4fabe96b3c", 00:26:51.906 "strip_size_kb": 64, 00:26:51.906 "state": "online", 00:26:51.906 "raid_level": "raid0", 00:26:51.906 "superblock": true, 00:26:51.906 "num_base_bdevs": 2, 00:26:51.906 "num_base_bdevs_discovered": 2, 00:26:51.906 "num_base_bdevs_operational": 2, 00:26:51.906 "base_bdevs_list": [ 00:26:51.906 { 00:26:51.906 "name": "BaseBdev1", 00:26:51.906 "uuid": "e2c3d671-3edd-5310-9bed-e3e7ae9d7b4b", 00:26:51.906 "is_configured": true, 00:26:51.906 "data_offset": 2048, 00:26:51.906 "data_size": 63488 00:26:51.906 }, 00:26:51.906 { 00:26:51.906 "name": "BaseBdev2", 00:26:51.906 "uuid": "083b9a96-46de-5bad-868b-6b6c6dfb2d80", 00:26:51.906 "is_configured": true, 00:26:51.906 "data_offset": 2048, 00:26:51.906 "data_size": 63488 00:26:51.906 } 00:26:51.906 ] 00:26:51.906 }' 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:51.906 13:46:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:52.167 13:46:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:26:52.167 13:46:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:26:52.425 [2024-11-20 13:46:59.254805] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:26:53.356 13:47:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:26:53.356 13:47:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:53.356 13:47:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:53.356 13:47:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:53.356 13:47:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:26:53.356 13:47:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:26:53.356 13:47:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:26:53.356 13:47:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:26:53.356 13:47:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:26:53.356 13:47:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:26:53.356 13:47:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:26:53.356 13:47:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:26:53.356 13:47:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:53.356 13:47:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:53.356 13:47:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:53.356 13:47:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:53.356 13:47:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:53.356 13:47:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:53.356 13:47:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:53.356 13:47:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:53.356 13:47:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:53.356 13:47:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:53.356 13:47:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:53.356 "name": "raid_bdev1", 00:26:53.356 "uuid": "1d5535f9-4a26-4f81-8c6e-5f4fabe96b3c", 00:26:53.356 "strip_size_kb": 64, 00:26:53.356 "state": "online", 00:26:53.356 "raid_level": "raid0", 00:26:53.356 "superblock": true, 00:26:53.356 "num_base_bdevs": 2, 00:26:53.356 "num_base_bdevs_discovered": 2, 00:26:53.356 "num_base_bdevs_operational": 2, 00:26:53.356 "base_bdevs_list": [ 00:26:53.356 { 00:26:53.356 "name": "BaseBdev1", 00:26:53.356 "uuid": "e2c3d671-3edd-5310-9bed-e3e7ae9d7b4b", 00:26:53.356 "is_configured": true, 00:26:53.356 "data_offset": 2048, 00:26:53.356 "data_size": 63488 00:26:53.356 }, 00:26:53.356 { 00:26:53.356 "name": "BaseBdev2", 00:26:53.356 "uuid": "083b9a96-46de-5bad-868b-6b6c6dfb2d80", 00:26:53.356 "is_configured": true, 00:26:53.356 "data_offset": 2048, 00:26:53.356 "data_size": 63488 00:26:53.356 } 00:26:53.356 ] 00:26:53.356 }' 00:26:53.356 13:47:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:53.356 13:47:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:53.614 13:47:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:26:53.614 13:47:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:53.614 13:47:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:53.614 [2024-11-20 13:47:00.541922] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:26:53.614 [2024-11-20 13:47:00.541959] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:26:53.614 [2024-11-20 13:47:00.545130] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:53.614 [2024-11-20 13:47:00.545180] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:53.614 [2024-11-20 13:47:00.545215] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:26:53.614 [2024-11-20 13:47:00.545227] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:26:53.614 { 00:26:53.614 "results": [ 00:26:53.614 { 00:26:53.614 "job": "raid_bdev1", 00:26:53.614 "core_mask": "0x1", 00:26:53.614 "workload": "randrw", 00:26:53.614 "percentage": 50, 00:26:53.614 "status": "finished", 00:26:53.614 "queue_depth": 1, 00:26:53.614 "io_size": 131072, 00:26:53.614 "runtime": 1.285052, 00:26:53.614 "iops": 13834.45961719837, 00:26:53.614 "mibps": 1729.3074521497963, 00:26:53.614 "io_failed": 1, 00:26:53.614 "io_timeout": 0, 00:26:53.614 "avg_latency_us": 98.90210940305546, 00:26:53.614 "min_latency_us": 33.870769230769234, 00:26:53.614 "max_latency_us": 1688.8123076923077 00:26:53.614 } 00:26:53.614 ], 00:26:53.614 "core_count": 1 00:26:53.614 } 00:26:53.614 13:47:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:53.614 13:47:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 60216 00:26:53.614 13:47:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 60216 ']' 00:26:53.614 13:47:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 60216 00:26:53.614 13:47:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:26:53.614 13:47:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:53.614 13:47:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 60216 00:26:53.614 killing process with pid 60216 00:26:53.614 13:47:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:26:53.614 13:47:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:26:53.614 13:47:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 60216' 00:26:53.614 13:47:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 60216 00:26:53.614 [2024-11-20 13:47:00.577323] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:26:53.614 13:47:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 60216 00:26:53.614 [2024-11-20 13:47:00.665792] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:26:54.548 13:47:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.4txNlrfn0o 00:26:54.548 13:47:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:26:54.548 13:47:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:26:54.548 13:47:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.78 00:26:54.548 13:47:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:26:54.548 13:47:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:26:54.548 13:47:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:26:54.548 13:47:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.78 != \0\.\0\0 ]] 00:26:54.548 00:26:54.548 real 0m3.679s 00:26:54.548 user 0m4.418s 00:26:54.548 sys 0m0.427s 00:26:54.548 ************************************ 00:26:54.548 END TEST raid_write_error_test 00:26:54.548 ************************************ 00:26:54.548 13:47:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:26:54.548 13:47:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:26:54.548 13:47:01 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:26:54.548 13:47:01 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 2 false 00:26:54.548 13:47:01 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:26:54.548 13:47:01 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:26:54.548 13:47:01 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:26:54.548 ************************************ 00:26:54.548 START TEST raid_state_function_test 00:26:54.548 ************************************ 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 2 false 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:26:54.548 Process raid pid: 60354 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=60354 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 60354' 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 60354 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 60354 ']' 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:54.548 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:54.548 13:47:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:54.549 13:47:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:26:54.549 13:47:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:54.549 [2024-11-20 13:47:01.576195] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:26:54.549 [2024-11-20 13:47:01.576303] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:54.807 [2024-11-20 13:47:01.726966] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:54.807 [2024-11-20 13:47:01.847904] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:55.064 [2024-11-20 13:47:01.999419] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:55.064 [2024-11-20 13:47:01.999655] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:55.631 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:55.631 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:26:55.631 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:26:55.631 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:55.631 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:55.631 [2024-11-20 13:47:02.450463] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:26:55.631 [2024-11-20 13:47:02.450530] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:26:55.631 [2024-11-20 13:47:02.450542] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:26:55.631 [2024-11-20 13:47:02.450552] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:26:55.631 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:55.631 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:26:55.631 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:26:55.631 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:26:55.631 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:26:55.631 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:26:55.631 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:55.631 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:55.631 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:55.631 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:55.631 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:55.631 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:26:55.631 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:55.631 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:55.631 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:55.631 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:55.631 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:55.631 "name": "Existed_Raid", 00:26:55.631 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:55.631 "strip_size_kb": 64, 00:26:55.631 "state": "configuring", 00:26:55.631 "raid_level": "concat", 00:26:55.631 "superblock": false, 00:26:55.631 "num_base_bdevs": 2, 00:26:55.631 "num_base_bdevs_discovered": 0, 00:26:55.631 "num_base_bdevs_operational": 2, 00:26:55.631 "base_bdevs_list": [ 00:26:55.631 { 00:26:55.631 "name": "BaseBdev1", 00:26:55.631 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:55.631 "is_configured": false, 00:26:55.631 "data_offset": 0, 00:26:55.631 "data_size": 0 00:26:55.631 }, 00:26:55.631 { 00:26:55.631 "name": "BaseBdev2", 00:26:55.631 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:55.631 "is_configured": false, 00:26:55.631 "data_offset": 0, 00:26:55.631 "data_size": 0 00:26:55.631 } 00:26:55.631 ] 00:26:55.631 }' 00:26:55.631 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:55.631 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:55.892 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:26:55.892 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:55.892 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:55.892 [2024-11-20 13:47:02.754478] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:26:55.892 [2024-11-20 13:47:02.754518] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:26:55.892 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:55.892 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:26:55.892 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:55.892 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:55.892 [2024-11-20 13:47:02.762456] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:26:55.892 [2024-11-20 13:47:02.762504] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:26:55.892 [2024-11-20 13:47:02.762515] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:26:55.892 [2024-11-20 13:47:02.762527] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:26:55.892 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:55.892 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:26:55.892 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:55.892 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:55.892 [2024-11-20 13:47:02.797908] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:26:55.892 BaseBdev1 00:26:55.892 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:55.892 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:26:55.892 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:26:55.892 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:26:55.892 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:26:55.892 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:26:55.892 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:26:55.892 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:26:55.892 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:55.892 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:55.892 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:55.892 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:26:55.892 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:55.892 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:55.892 [ 00:26:55.892 { 00:26:55.892 "name": "BaseBdev1", 00:26:55.892 "aliases": [ 00:26:55.892 "ca8ef170-3793-4d0c-b17c-006604dbd75c" 00:26:55.892 ], 00:26:55.892 "product_name": "Malloc disk", 00:26:55.892 "block_size": 512, 00:26:55.892 "num_blocks": 65536, 00:26:55.892 "uuid": "ca8ef170-3793-4d0c-b17c-006604dbd75c", 00:26:55.892 "assigned_rate_limits": { 00:26:55.892 "rw_ios_per_sec": 0, 00:26:55.892 "rw_mbytes_per_sec": 0, 00:26:55.892 "r_mbytes_per_sec": 0, 00:26:55.892 "w_mbytes_per_sec": 0 00:26:55.892 }, 00:26:55.892 "claimed": true, 00:26:55.892 "claim_type": "exclusive_write", 00:26:55.892 "zoned": false, 00:26:55.892 "supported_io_types": { 00:26:55.892 "read": true, 00:26:55.892 "write": true, 00:26:55.892 "unmap": true, 00:26:55.892 "flush": true, 00:26:55.892 "reset": true, 00:26:55.892 "nvme_admin": false, 00:26:55.892 "nvme_io": false, 00:26:55.892 "nvme_io_md": false, 00:26:55.892 "write_zeroes": true, 00:26:55.892 "zcopy": true, 00:26:55.893 "get_zone_info": false, 00:26:55.893 "zone_management": false, 00:26:55.893 "zone_append": false, 00:26:55.893 "compare": false, 00:26:55.893 "compare_and_write": false, 00:26:55.893 "abort": true, 00:26:55.893 "seek_hole": false, 00:26:55.893 "seek_data": false, 00:26:55.893 "copy": true, 00:26:55.893 "nvme_iov_md": false 00:26:55.893 }, 00:26:55.893 "memory_domains": [ 00:26:55.893 { 00:26:55.893 "dma_device_id": "system", 00:26:55.893 "dma_device_type": 1 00:26:55.893 }, 00:26:55.893 { 00:26:55.893 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:55.893 "dma_device_type": 2 00:26:55.893 } 00:26:55.893 ], 00:26:55.893 "driver_specific": {} 00:26:55.893 } 00:26:55.893 ] 00:26:55.893 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:55.893 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:26:55.893 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:26:55.893 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:26:55.893 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:26:55.893 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:26:55.893 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:26:55.893 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:55.893 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:55.893 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:55.893 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:55.893 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:55.893 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:55.893 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:55.893 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:55.893 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:26:55.893 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:55.893 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:55.893 "name": "Existed_Raid", 00:26:55.893 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:55.893 "strip_size_kb": 64, 00:26:55.893 "state": "configuring", 00:26:55.893 "raid_level": "concat", 00:26:55.893 "superblock": false, 00:26:55.893 "num_base_bdevs": 2, 00:26:55.893 "num_base_bdevs_discovered": 1, 00:26:55.893 "num_base_bdevs_operational": 2, 00:26:55.893 "base_bdevs_list": [ 00:26:55.893 { 00:26:55.893 "name": "BaseBdev1", 00:26:55.893 "uuid": "ca8ef170-3793-4d0c-b17c-006604dbd75c", 00:26:55.893 "is_configured": true, 00:26:55.893 "data_offset": 0, 00:26:55.893 "data_size": 65536 00:26:55.893 }, 00:26:55.893 { 00:26:55.893 "name": "BaseBdev2", 00:26:55.893 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:55.893 "is_configured": false, 00:26:55.893 "data_offset": 0, 00:26:55.893 "data_size": 0 00:26:55.893 } 00:26:55.893 ] 00:26:55.893 }' 00:26:55.893 13:47:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:55.893 13:47:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:56.152 [2024-11-20 13:47:03.146038] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:26:56.152 [2024-11-20 13:47:03.146192] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:56.152 [2024-11-20 13:47:03.154080] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:26:56.152 [2024-11-20 13:47:03.156133] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:26:56.152 [2024-11-20 13:47:03.156251] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:56.152 "name": "Existed_Raid", 00:26:56.152 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:56.152 "strip_size_kb": 64, 00:26:56.152 "state": "configuring", 00:26:56.152 "raid_level": "concat", 00:26:56.152 "superblock": false, 00:26:56.152 "num_base_bdevs": 2, 00:26:56.152 "num_base_bdevs_discovered": 1, 00:26:56.152 "num_base_bdevs_operational": 2, 00:26:56.152 "base_bdevs_list": [ 00:26:56.152 { 00:26:56.152 "name": "BaseBdev1", 00:26:56.152 "uuid": "ca8ef170-3793-4d0c-b17c-006604dbd75c", 00:26:56.152 "is_configured": true, 00:26:56.152 "data_offset": 0, 00:26:56.152 "data_size": 65536 00:26:56.152 }, 00:26:56.152 { 00:26:56.152 "name": "BaseBdev2", 00:26:56.152 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:56.152 "is_configured": false, 00:26:56.152 "data_offset": 0, 00:26:56.152 "data_size": 0 00:26:56.152 } 00:26:56.152 ] 00:26:56.152 }' 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:56.152 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:56.411 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:26:56.411 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:56.411 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:56.411 [2024-11-20 13:47:03.458670] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:26:56.411 [2024-11-20 13:47:03.458718] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:26:56.411 [2024-11-20 13:47:03.458725] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:26:56.411 [2024-11-20 13:47:03.458957] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:26:56.411 [2024-11-20 13:47:03.459153] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:26:56.411 [2024-11-20 13:47:03.459164] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:26:56.411 [2024-11-20 13:47:03.459415] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:56.411 BaseBdev2 00:26:56.411 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:56.411 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:26:56.411 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:26:56.411 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:26:56.411 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:26:56.411 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:26:56.411 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:26:56.411 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:26:56.411 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:56.411 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:56.668 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:56.668 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:26:56.668 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:56.668 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:56.668 [ 00:26:56.668 { 00:26:56.668 "name": "BaseBdev2", 00:26:56.668 "aliases": [ 00:26:56.668 "acb1f243-107a-49d6-ad3a-1bcaaee51c24" 00:26:56.668 ], 00:26:56.668 "product_name": "Malloc disk", 00:26:56.668 "block_size": 512, 00:26:56.668 "num_blocks": 65536, 00:26:56.668 "uuid": "acb1f243-107a-49d6-ad3a-1bcaaee51c24", 00:26:56.668 "assigned_rate_limits": { 00:26:56.668 "rw_ios_per_sec": 0, 00:26:56.668 "rw_mbytes_per_sec": 0, 00:26:56.668 "r_mbytes_per_sec": 0, 00:26:56.668 "w_mbytes_per_sec": 0 00:26:56.668 }, 00:26:56.668 "claimed": true, 00:26:56.668 "claim_type": "exclusive_write", 00:26:56.668 "zoned": false, 00:26:56.668 "supported_io_types": { 00:26:56.668 "read": true, 00:26:56.668 "write": true, 00:26:56.668 "unmap": true, 00:26:56.668 "flush": true, 00:26:56.668 "reset": true, 00:26:56.668 "nvme_admin": false, 00:26:56.668 "nvme_io": false, 00:26:56.668 "nvme_io_md": false, 00:26:56.668 "write_zeroes": true, 00:26:56.668 "zcopy": true, 00:26:56.668 "get_zone_info": false, 00:26:56.668 "zone_management": false, 00:26:56.669 "zone_append": false, 00:26:56.669 "compare": false, 00:26:56.669 "compare_and_write": false, 00:26:56.669 "abort": true, 00:26:56.669 "seek_hole": false, 00:26:56.669 "seek_data": false, 00:26:56.669 "copy": true, 00:26:56.669 "nvme_iov_md": false 00:26:56.669 }, 00:26:56.669 "memory_domains": [ 00:26:56.669 { 00:26:56.669 "dma_device_id": "system", 00:26:56.669 "dma_device_type": 1 00:26:56.669 }, 00:26:56.669 { 00:26:56.669 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:56.669 "dma_device_type": 2 00:26:56.669 } 00:26:56.669 ], 00:26:56.669 "driver_specific": {} 00:26:56.669 } 00:26:56.669 ] 00:26:56.669 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:56.669 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:26:56.669 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:26:56.669 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:26:56.669 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 2 00:26:56.669 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:26:56.669 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:26:56.669 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:26:56.669 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:26:56.669 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:56.669 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:56.669 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:56.669 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:56.669 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:56.669 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:56.669 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:26:56.669 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:56.669 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:56.669 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:56.669 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:56.669 "name": "Existed_Raid", 00:26:56.669 "uuid": "482bccf9-8bc9-4b15-989b-b73e3ec3f213", 00:26:56.669 "strip_size_kb": 64, 00:26:56.669 "state": "online", 00:26:56.669 "raid_level": "concat", 00:26:56.669 "superblock": false, 00:26:56.669 "num_base_bdevs": 2, 00:26:56.669 "num_base_bdevs_discovered": 2, 00:26:56.669 "num_base_bdevs_operational": 2, 00:26:56.669 "base_bdevs_list": [ 00:26:56.669 { 00:26:56.669 "name": "BaseBdev1", 00:26:56.669 "uuid": "ca8ef170-3793-4d0c-b17c-006604dbd75c", 00:26:56.669 "is_configured": true, 00:26:56.669 "data_offset": 0, 00:26:56.669 "data_size": 65536 00:26:56.669 }, 00:26:56.669 { 00:26:56.669 "name": "BaseBdev2", 00:26:56.669 "uuid": "acb1f243-107a-49d6-ad3a-1bcaaee51c24", 00:26:56.669 "is_configured": true, 00:26:56.669 "data_offset": 0, 00:26:56.669 "data_size": 65536 00:26:56.669 } 00:26:56.669 ] 00:26:56.669 }' 00:26:56.669 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:56.669 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:56.927 [2024-11-20 13:47:03.811052] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:26:56.927 "name": "Existed_Raid", 00:26:56.927 "aliases": [ 00:26:56.927 "482bccf9-8bc9-4b15-989b-b73e3ec3f213" 00:26:56.927 ], 00:26:56.927 "product_name": "Raid Volume", 00:26:56.927 "block_size": 512, 00:26:56.927 "num_blocks": 131072, 00:26:56.927 "uuid": "482bccf9-8bc9-4b15-989b-b73e3ec3f213", 00:26:56.927 "assigned_rate_limits": { 00:26:56.927 "rw_ios_per_sec": 0, 00:26:56.927 "rw_mbytes_per_sec": 0, 00:26:56.927 "r_mbytes_per_sec": 0, 00:26:56.927 "w_mbytes_per_sec": 0 00:26:56.927 }, 00:26:56.927 "claimed": false, 00:26:56.927 "zoned": false, 00:26:56.927 "supported_io_types": { 00:26:56.927 "read": true, 00:26:56.927 "write": true, 00:26:56.927 "unmap": true, 00:26:56.927 "flush": true, 00:26:56.927 "reset": true, 00:26:56.927 "nvme_admin": false, 00:26:56.927 "nvme_io": false, 00:26:56.927 "nvme_io_md": false, 00:26:56.927 "write_zeroes": true, 00:26:56.927 "zcopy": false, 00:26:56.927 "get_zone_info": false, 00:26:56.927 "zone_management": false, 00:26:56.927 "zone_append": false, 00:26:56.927 "compare": false, 00:26:56.927 "compare_and_write": false, 00:26:56.927 "abort": false, 00:26:56.927 "seek_hole": false, 00:26:56.927 "seek_data": false, 00:26:56.927 "copy": false, 00:26:56.927 "nvme_iov_md": false 00:26:56.927 }, 00:26:56.927 "memory_domains": [ 00:26:56.927 { 00:26:56.927 "dma_device_id": "system", 00:26:56.927 "dma_device_type": 1 00:26:56.927 }, 00:26:56.927 { 00:26:56.927 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:56.927 "dma_device_type": 2 00:26:56.927 }, 00:26:56.927 { 00:26:56.927 "dma_device_id": "system", 00:26:56.927 "dma_device_type": 1 00:26:56.927 }, 00:26:56.927 { 00:26:56.927 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:56.927 "dma_device_type": 2 00:26:56.927 } 00:26:56.927 ], 00:26:56.927 "driver_specific": { 00:26:56.927 "raid": { 00:26:56.927 "uuid": "482bccf9-8bc9-4b15-989b-b73e3ec3f213", 00:26:56.927 "strip_size_kb": 64, 00:26:56.927 "state": "online", 00:26:56.927 "raid_level": "concat", 00:26:56.927 "superblock": false, 00:26:56.927 "num_base_bdevs": 2, 00:26:56.927 "num_base_bdevs_discovered": 2, 00:26:56.927 "num_base_bdevs_operational": 2, 00:26:56.927 "base_bdevs_list": [ 00:26:56.927 { 00:26:56.927 "name": "BaseBdev1", 00:26:56.927 "uuid": "ca8ef170-3793-4d0c-b17c-006604dbd75c", 00:26:56.927 "is_configured": true, 00:26:56.927 "data_offset": 0, 00:26:56.927 "data_size": 65536 00:26:56.927 }, 00:26:56.927 { 00:26:56.927 "name": "BaseBdev2", 00:26:56.927 "uuid": "acb1f243-107a-49d6-ad3a-1bcaaee51c24", 00:26:56.927 "is_configured": true, 00:26:56.927 "data_offset": 0, 00:26:56.927 "data_size": 65536 00:26:56.927 } 00:26:56.927 ] 00:26:56.927 } 00:26:56.927 } 00:26:56.927 }' 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:26:56.927 BaseBdev2' 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:56.927 13:47:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:56.927 [2024-11-20 13:47:03.966848] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:26:56.927 [2024-11-20 13:47:03.966881] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:26:56.927 [2024-11-20 13:47:03.966928] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:57.185 13:47:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:57.185 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:26:57.185 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:26:57.185 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:26:57.185 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:26:57.185 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:26:57.185 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 1 00:26:57.185 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:26:57.185 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:26:57.185 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:26:57.185 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:26:57.185 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:26:57.185 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:57.185 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:57.185 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:57.185 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:57.185 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:26:57.185 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:57.185 13:47:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:57.185 13:47:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:57.185 13:47:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:57.185 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:57.185 "name": "Existed_Raid", 00:26:57.185 "uuid": "482bccf9-8bc9-4b15-989b-b73e3ec3f213", 00:26:57.185 "strip_size_kb": 64, 00:26:57.185 "state": "offline", 00:26:57.185 "raid_level": "concat", 00:26:57.185 "superblock": false, 00:26:57.185 "num_base_bdevs": 2, 00:26:57.185 "num_base_bdevs_discovered": 1, 00:26:57.185 "num_base_bdevs_operational": 1, 00:26:57.185 "base_bdevs_list": [ 00:26:57.185 { 00:26:57.185 "name": null, 00:26:57.185 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:57.185 "is_configured": false, 00:26:57.185 "data_offset": 0, 00:26:57.185 "data_size": 65536 00:26:57.185 }, 00:26:57.185 { 00:26:57.185 "name": "BaseBdev2", 00:26:57.186 "uuid": "acb1f243-107a-49d6-ad3a-1bcaaee51c24", 00:26:57.186 "is_configured": true, 00:26:57.186 "data_offset": 0, 00:26:57.186 "data_size": 65536 00:26:57.186 } 00:26:57.186 ] 00:26:57.186 }' 00:26:57.186 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:57.186 13:47:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:57.444 [2024-11-20 13:47:04.360681] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:26:57.444 [2024-11-20 13:47:04.360744] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 60354 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 60354 ']' 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 60354 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 60354 00:26:57.444 killing process with pid 60354 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 60354' 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 60354 00:26:57.444 [2024-11-20 13:47:04.468808] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:26:57.444 13:47:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 60354 00:26:57.444 [2024-11-20 13:47:04.477838] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:26:58.380 00:26:58.380 real 0m3.594s 00:26:58.380 user 0m5.166s 00:26:58.380 sys 0m0.610s 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:26:58.380 ************************************ 00:26:58.380 END TEST raid_state_function_test 00:26:58.380 ************************************ 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:26:58.380 13:47:05 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 2 true 00:26:58.380 13:47:05 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:26:58.380 13:47:05 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:26:58.380 13:47:05 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:26:58.380 ************************************ 00:26:58.380 START TEST raid_state_function_test_sb 00:26:58.380 ************************************ 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 2 true 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:26:58.380 Process raid pid: 60588 00:26:58.380 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=60588 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 60588' 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 60588 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 60588 ']' 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:58.380 13:47:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:26:58.380 [2024-11-20 13:47:05.223475] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:26:58.380 [2024-11-20 13:47:05.223789] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:58.380 [2024-11-20 13:47:05.383837] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:58.638 [2024-11-20 13:47:05.502613] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:58.638 [2024-11-20 13:47:05.652060] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:58.638 [2024-11-20 13:47:05.652274] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:59.208 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:59.208 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:26:59.208 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:26:59.208 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:59.208 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:59.208 [2024-11-20 13:47:06.081299] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:26:59.208 [2024-11-20 13:47:06.081360] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:26:59.208 [2024-11-20 13:47:06.081371] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:26:59.208 [2024-11-20 13:47:06.081381] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:26:59.208 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:59.208 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:26:59.208 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:26:59.208 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:26:59.208 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:26:59.208 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:26:59.208 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:59.208 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:59.208 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:59.208 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:59.208 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:59.208 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:59.208 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:59.208 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:59.208 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:26:59.208 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:59.208 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:59.208 "name": "Existed_Raid", 00:26:59.208 "uuid": "ef9bee42-3454-47b6-9520-a6056a1cd33d", 00:26:59.208 "strip_size_kb": 64, 00:26:59.208 "state": "configuring", 00:26:59.208 "raid_level": "concat", 00:26:59.208 "superblock": true, 00:26:59.208 "num_base_bdevs": 2, 00:26:59.208 "num_base_bdevs_discovered": 0, 00:26:59.208 "num_base_bdevs_operational": 2, 00:26:59.208 "base_bdevs_list": [ 00:26:59.208 { 00:26:59.208 "name": "BaseBdev1", 00:26:59.208 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:59.208 "is_configured": false, 00:26:59.208 "data_offset": 0, 00:26:59.208 "data_size": 0 00:26:59.208 }, 00:26:59.208 { 00:26:59.208 "name": "BaseBdev2", 00:26:59.208 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:59.208 "is_configured": false, 00:26:59.208 "data_offset": 0, 00:26:59.208 "data_size": 0 00:26:59.208 } 00:26:59.208 ] 00:26:59.208 }' 00:26:59.208 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:59.208 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:59.466 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:26:59.466 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:59.466 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:59.466 [2024-11-20 13:47:06.381296] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:26:59.466 [2024-11-20 13:47:06.381334] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:26:59.466 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:59.466 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:26:59.466 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:59.466 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:59.466 [2024-11-20 13:47:06.389287] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:26:59.466 [2024-11-20 13:47:06.389328] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:26:59.466 [2024-11-20 13:47:06.389338] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:26:59.466 [2024-11-20 13:47:06.389350] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:26:59.466 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:59.466 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:26:59.466 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:59.467 [2024-11-20 13:47:06.424051] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:26:59.467 BaseBdev1 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:59.467 [ 00:26:59.467 { 00:26:59.467 "name": "BaseBdev1", 00:26:59.467 "aliases": [ 00:26:59.467 "1ffe861d-038b-4381-901c-6377803c8893" 00:26:59.467 ], 00:26:59.467 "product_name": "Malloc disk", 00:26:59.467 "block_size": 512, 00:26:59.467 "num_blocks": 65536, 00:26:59.467 "uuid": "1ffe861d-038b-4381-901c-6377803c8893", 00:26:59.467 "assigned_rate_limits": { 00:26:59.467 "rw_ios_per_sec": 0, 00:26:59.467 "rw_mbytes_per_sec": 0, 00:26:59.467 "r_mbytes_per_sec": 0, 00:26:59.467 "w_mbytes_per_sec": 0 00:26:59.467 }, 00:26:59.467 "claimed": true, 00:26:59.467 "claim_type": "exclusive_write", 00:26:59.467 "zoned": false, 00:26:59.467 "supported_io_types": { 00:26:59.467 "read": true, 00:26:59.467 "write": true, 00:26:59.467 "unmap": true, 00:26:59.467 "flush": true, 00:26:59.467 "reset": true, 00:26:59.467 "nvme_admin": false, 00:26:59.467 "nvme_io": false, 00:26:59.467 "nvme_io_md": false, 00:26:59.467 "write_zeroes": true, 00:26:59.467 "zcopy": true, 00:26:59.467 "get_zone_info": false, 00:26:59.467 "zone_management": false, 00:26:59.467 "zone_append": false, 00:26:59.467 "compare": false, 00:26:59.467 "compare_and_write": false, 00:26:59.467 "abort": true, 00:26:59.467 "seek_hole": false, 00:26:59.467 "seek_data": false, 00:26:59.467 "copy": true, 00:26:59.467 "nvme_iov_md": false 00:26:59.467 }, 00:26:59.467 "memory_domains": [ 00:26:59.467 { 00:26:59.467 "dma_device_id": "system", 00:26:59.467 "dma_device_type": 1 00:26:59.467 }, 00:26:59.467 { 00:26:59.467 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:26:59.467 "dma_device_type": 2 00:26:59.467 } 00:26:59.467 ], 00:26:59.467 "driver_specific": {} 00:26:59.467 } 00:26:59.467 ] 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:59.467 "name": "Existed_Raid", 00:26:59.467 "uuid": "f22b9a60-2195-45f6-9d4a-7949efa13c54", 00:26:59.467 "strip_size_kb": 64, 00:26:59.467 "state": "configuring", 00:26:59.467 "raid_level": "concat", 00:26:59.467 "superblock": true, 00:26:59.467 "num_base_bdevs": 2, 00:26:59.467 "num_base_bdevs_discovered": 1, 00:26:59.467 "num_base_bdevs_operational": 2, 00:26:59.467 "base_bdevs_list": [ 00:26:59.467 { 00:26:59.467 "name": "BaseBdev1", 00:26:59.467 "uuid": "1ffe861d-038b-4381-901c-6377803c8893", 00:26:59.467 "is_configured": true, 00:26:59.467 "data_offset": 2048, 00:26:59.467 "data_size": 63488 00:26:59.467 }, 00:26:59.467 { 00:26:59.467 "name": "BaseBdev2", 00:26:59.467 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:59.467 "is_configured": false, 00:26:59.467 "data_offset": 0, 00:26:59.467 "data_size": 0 00:26:59.467 } 00:26:59.467 ] 00:26:59.467 }' 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:59.467 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:59.794 [2024-11-20 13:47:06.744166] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:26:59.794 [2024-11-20 13:47:06.744221] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:59.794 [2024-11-20 13:47:06.752222] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:26:59.794 [2024-11-20 13:47:06.754207] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:26:59.794 [2024-11-20 13:47:06.754248] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:59.794 "name": "Existed_Raid", 00:26:59.794 "uuid": "dbfbeb2b-29a8-4909-95b8-dbd216c668d1", 00:26:59.794 "strip_size_kb": 64, 00:26:59.794 "state": "configuring", 00:26:59.794 "raid_level": "concat", 00:26:59.794 "superblock": true, 00:26:59.794 "num_base_bdevs": 2, 00:26:59.794 "num_base_bdevs_discovered": 1, 00:26:59.794 "num_base_bdevs_operational": 2, 00:26:59.794 "base_bdevs_list": [ 00:26:59.794 { 00:26:59.794 "name": "BaseBdev1", 00:26:59.794 "uuid": "1ffe861d-038b-4381-901c-6377803c8893", 00:26:59.794 "is_configured": true, 00:26:59.794 "data_offset": 2048, 00:26:59.794 "data_size": 63488 00:26:59.794 }, 00:26:59.794 { 00:26:59.794 "name": "BaseBdev2", 00:26:59.794 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:59.794 "is_configured": false, 00:26:59.794 "data_offset": 0, 00:26:59.794 "data_size": 0 00:26:59.794 } 00:26:59.794 ] 00:26:59.794 }' 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:59.794 13:47:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:00.070 [2024-11-20 13:47:07.084741] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:27:00.070 [2024-11-20 13:47:07.085015] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:27:00.070 [2024-11-20 13:47:07.085029] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:27:00.070 BaseBdev2 00:27:00.070 [2024-11-20 13:47:07.085306] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:27:00.070 [2024-11-20 13:47:07.085453] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:27:00.070 [2024-11-20 13:47:07.085465] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:27:00.070 [2024-11-20 13:47:07.085592] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:00.070 [ 00:27:00.070 { 00:27:00.070 "name": "BaseBdev2", 00:27:00.070 "aliases": [ 00:27:00.070 "dfd5a8f4-ae98-4763-ae45-086bd1046285" 00:27:00.070 ], 00:27:00.070 "product_name": "Malloc disk", 00:27:00.070 "block_size": 512, 00:27:00.070 "num_blocks": 65536, 00:27:00.070 "uuid": "dfd5a8f4-ae98-4763-ae45-086bd1046285", 00:27:00.070 "assigned_rate_limits": { 00:27:00.070 "rw_ios_per_sec": 0, 00:27:00.070 "rw_mbytes_per_sec": 0, 00:27:00.070 "r_mbytes_per_sec": 0, 00:27:00.070 "w_mbytes_per_sec": 0 00:27:00.070 }, 00:27:00.070 "claimed": true, 00:27:00.070 "claim_type": "exclusive_write", 00:27:00.070 "zoned": false, 00:27:00.070 "supported_io_types": { 00:27:00.070 "read": true, 00:27:00.070 "write": true, 00:27:00.070 "unmap": true, 00:27:00.070 "flush": true, 00:27:00.070 "reset": true, 00:27:00.070 "nvme_admin": false, 00:27:00.070 "nvme_io": false, 00:27:00.070 "nvme_io_md": false, 00:27:00.070 "write_zeroes": true, 00:27:00.070 "zcopy": true, 00:27:00.070 "get_zone_info": false, 00:27:00.070 "zone_management": false, 00:27:00.070 "zone_append": false, 00:27:00.070 "compare": false, 00:27:00.070 "compare_and_write": false, 00:27:00.070 "abort": true, 00:27:00.070 "seek_hole": false, 00:27:00.070 "seek_data": false, 00:27:00.070 "copy": true, 00:27:00.070 "nvme_iov_md": false 00:27:00.070 }, 00:27:00.070 "memory_domains": [ 00:27:00.070 { 00:27:00.070 "dma_device_id": "system", 00:27:00.070 "dma_device_type": 1 00:27:00.070 }, 00:27:00.070 { 00:27:00.070 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:00.070 "dma_device_type": 2 00:27:00.070 } 00:27:00.070 ], 00:27:00.070 "driver_specific": {} 00:27:00.070 } 00:27:00.070 ] 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 2 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:00.070 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:00.333 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:00.333 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:00.333 "name": "Existed_Raid", 00:27:00.333 "uuid": "dbfbeb2b-29a8-4909-95b8-dbd216c668d1", 00:27:00.333 "strip_size_kb": 64, 00:27:00.333 "state": "online", 00:27:00.333 "raid_level": "concat", 00:27:00.333 "superblock": true, 00:27:00.333 "num_base_bdevs": 2, 00:27:00.333 "num_base_bdevs_discovered": 2, 00:27:00.333 "num_base_bdevs_operational": 2, 00:27:00.333 "base_bdevs_list": [ 00:27:00.333 { 00:27:00.333 "name": "BaseBdev1", 00:27:00.333 "uuid": "1ffe861d-038b-4381-901c-6377803c8893", 00:27:00.333 "is_configured": true, 00:27:00.333 "data_offset": 2048, 00:27:00.333 "data_size": 63488 00:27:00.333 }, 00:27:00.333 { 00:27:00.333 "name": "BaseBdev2", 00:27:00.333 "uuid": "dfd5a8f4-ae98-4763-ae45-086bd1046285", 00:27:00.333 "is_configured": true, 00:27:00.333 "data_offset": 2048, 00:27:00.333 "data_size": 63488 00:27:00.333 } 00:27:00.333 ] 00:27:00.333 }' 00:27:00.333 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:00.333 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:00.592 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:27:00.592 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:27:00.592 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:27:00.592 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:27:00.592 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:27:00.592 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:27:00.592 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:27:00.592 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:27:00.592 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:00.592 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:00.592 [2024-11-20 13:47:07.441195] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:00.592 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:00.592 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:27:00.592 "name": "Existed_Raid", 00:27:00.592 "aliases": [ 00:27:00.592 "dbfbeb2b-29a8-4909-95b8-dbd216c668d1" 00:27:00.592 ], 00:27:00.592 "product_name": "Raid Volume", 00:27:00.592 "block_size": 512, 00:27:00.592 "num_blocks": 126976, 00:27:00.592 "uuid": "dbfbeb2b-29a8-4909-95b8-dbd216c668d1", 00:27:00.592 "assigned_rate_limits": { 00:27:00.592 "rw_ios_per_sec": 0, 00:27:00.592 "rw_mbytes_per_sec": 0, 00:27:00.592 "r_mbytes_per_sec": 0, 00:27:00.592 "w_mbytes_per_sec": 0 00:27:00.592 }, 00:27:00.592 "claimed": false, 00:27:00.592 "zoned": false, 00:27:00.592 "supported_io_types": { 00:27:00.592 "read": true, 00:27:00.592 "write": true, 00:27:00.592 "unmap": true, 00:27:00.592 "flush": true, 00:27:00.592 "reset": true, 00:27:00.592 "nvme_admin": false, 00:27:00.592 "nvme_io": false, 00:27:00.592 "nvme_io_md": false, 00:27:00.592 "write_zeroes": true, 00:27:00.592 "zcopy": false, 00:27:00.592 "get_zone_info": false, 00:27:00.592 "zone_management": false, 00:27:00.592 "zone_append": false, 00:27:00.592 "compare": false, 00:27:00.592 "compare_and_write": false, 00:27:00.592 "abort": false, 00:27:00.592 "seek_hole": false, 00:27:00.592 "seek_data": false, 00:27:00.592 "copy": false, 00:27:00.592 "nvme_iov_md": false 00:27:00.592 }, 00:27:00.592 "memory_domains": [ 00:27:00.592 { 00:27:00.592 "dma_device_id": "system", 00:27:00.592 "dma_device_type": 1 00:27:00.592 }, 00:27:00.592 { 00:27:00.592 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:00.592 "dma_device_type": 2 00:27:00.592 }, 00:27:00.592 { 00:27:00.592 "dma_device_id": "system", 00:27:00.592 "dma_device_type": 1 00:27:00.592 }, 00:27:00.592 { 00:27:00.592 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:00.592 "dma_device_type": 2 00:27:00.592 } 00:27:00.592 ], 00:27:00.592 "driver_specific": { 00:27:00.592 "raid": { 00:27:00.592 "uuid": "dbfbeb2b-29a8-4909-95b8-dbd216c668d1", 00:27:00.592 "strip_size_kb": 64, 00:27:00.592 "state": "online", 00:27:00.592 "raid_level": "concat", 00:27:00.592 "superblock": true, 00:27:00.592 "num_base_bdevs": 2, 00:27:00.592 "num_base_bdevs_discovered": 2, 00:27:00.592 "num_base_bdevs_operational": 2, 00:27:00.592 "base_bdevs_list": [ 00:27:00.592 { 00:27:00.592 "name": "BaseBdev1", 00:27:00.592 "uuid": "1ffe861d-038b-4381-901c-6377803c8893", 00:27:00.592 "is_configured": true, 00:27:00.592 "data_offset": 2048, 00:27:00.592 "data_size": 63488 00:27:00.592 }, 00:27:00.592 { 00:27:00.592 "name": "BaseBdev2", 00:27:00.592 "uuid": "dfd5a8f4-ae98-4763-ae45-086bd1046285", 00:27:00.592 "is_configured": true, 00:27:00.592 "data_offset": 2048, 00:27:00.592 "data_size": 63488 00:27:00.592 } 00:27:00.592 ] 00:27:00.592 } 00:27:00.592 } 00:27:00.592 }' 00:27:00.592 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:27:00.592 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:27:00.592 BaseBdev2' 00:27:00.592 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:00.592 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:27:00.592 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:00.592 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:00.592 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:27:00.592 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:00.593 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:00.593 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:00.593 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:00.593 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:00.593 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:00.593 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:27:00.593 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:00.593 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:00.593 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:00.593 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:00.593 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:00.593 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:00.593 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:27:00.593 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:00.593 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:00.593 [2024-11-20 13:47:07.600969] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:27:00.593 [2024-11-20 13:47:07.601018] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:27:00.593 [2024-11-20 13:47:07.601072] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:00.851 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:00.851 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:27:00.851 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:27:00.851 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:27:00.851 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:27:00.851 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:27:00.851 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 1 00:27:00.851 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:00.851 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:27:00.851 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:27:00.851 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:00.851 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:27:00.851 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:00.851 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:00.851 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:00.851 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:00.851 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:00.851 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:00.851 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:00.851 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:00.851 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:00.851 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:00.851 "name": "Existed_Raid", 00:27:00.851 "uuid": "dbfbeb2b-29a8-4909-95b8-dbd216c668d1", 00:27:00.851 "strip_size_kb": 64, 00:27:00.851 "state": "offline", 00:27:00.851 "raid_level": "concat", 00:27:00.851 "superblock": true, 00:27:00.851 "num_base_bdevs": 2, 00:27:00.851 "num_base_bdevs_discovered": 1, 00:27:00.851 "num_base_bdevs_operational": 1, 00:27:00.851 "base_bdevs_list": [ 00:27:00.851 { 00:27:00.851 "name": null, 00:27:00.851 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:00.851 "is_configured": false, 00:27:00.851 "data_offset": 0, 00:27:00.851 "data_size": 63488 00:27:00.851 }, 00:27:00.851 { 00:27:00.851 "name": "BaseBdev2", 00:27:00.851 "uuid": "dfd5a8f4-ae98-4763-ae45-086bd1046285", 00:27:00.851 "is_configured": true, 00:27:00.851 "data_offset": 2048, 00:27:00.851 "data_size": 63488 00:27:00.851 } 00:27:00.851 ] 00:27:00.851 }' 00:27:00.851 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:00.851 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:01.109 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:27:01.109 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:27:01.109 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:01.109 13:47:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:27:01.109 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:01.109 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:01.109 13:47:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:01.109 [2024-11-20 13:47:08.015428] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:27:01.109 [2024-11-20 13:47:08.015490] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 60588 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 60588 ']' 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 60588 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 60588 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:27:01.109 killing process with pid 60588 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 60588' 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 60588 00:27:01.109 [2024-11-20 13:47:08.144346] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:27:01.109 13:47:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 60588 00:27:01.109 [2024-11-20 13:47:08.155497] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:27:02.043 13:47:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:27:02.043 00:27:02.043 real 0m3.757s 00:27:02.043 user 0m5.318s 00:27:02.043 sys 0m0.638s 00:27:02.043 ************************************ 00:27:02.043 END TEST raid_state_function_test_sb 00:27:02.043 ************************************ 00:27:02.043 13:47:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:27:02.043 13:47:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:02.043 13:47:08 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 2 00:27:02.043 13:47:08 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:27:02.043 13:47:08 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:27:02.043 13:47:08 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:27:02.043 ************************************ 00:27:02.043 START TEST raid_superblock_test 00:27:02.043 ************************************ 00:27:02.043 13:47:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test concat 2 00:27:02.043 13:47:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:27:02.043 13:47:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:27:02.043 13:47:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:27:02.043 13:47:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:27:02.043 13:47:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:27:02.044 13:47:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:27:02.044 13:47:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:27:02.044 13:47:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:27:02.044 13:47:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:27:02.044 13:47:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:27:02.044 13:47:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:27:02.044 13:47:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:27:02.044 13:47:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:27:02.044 13:47:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:27:02.044 13:47:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:27:02.044 13:47:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:27:02.044 13:47:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=60829 00:27:02.044 13:47:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 60829 00:27:02.044 13:47:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 60829 ']' 00:27:02.044 13:47:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:02.044 13:47:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:02.044 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:02.044 13:47:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:02.044 13:47:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:02.044 13:47:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:02.044 13:47:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:27:02.044 [2024-11-20 13:47:09.022927] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:27:02.044 [2024-11-20 13:47:09.023041] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid60829 ] 00:27:02.318 [2024-11-20 13:47:09.178743] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:02.318 [2024-11-20 13:47:09.295081] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:27:02.583 [2024-11-20 13:47:09.443623] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:02.583 [2024-11-20 13:47:09.443680] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:02.840 13:47:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:02.840 13:47:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:27:02.840 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:27:02.840 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:27:03.098 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:27:03.098 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:03.099 malloc1 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:03.099 [2024-11-20 13:47:09.937304] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:27:03.099 [2024-11-20 13:47:09.937369] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:03.099 [2024-11-20 13:47:09.937392] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:27:03.099 [2024-11-20 13:47:09.937402] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:03.099 [2024-11-20 13:47:09.939667] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:03.099 [2024-11-20 13:47:09.939702] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:27:03.099 pt1 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:03.099 malloc2 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:03.099 [2024-11-20 13:47:09.979312] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:27:03.099 [2024-11-20 13:47:09.979368] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:03.099 [2024-11-20 13:47:09.979399] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:27:03.099 [2024-11-20 13:47:09.979408] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:03.099 [2024-11-20 13:47:09.981629] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:03.099 [2024-11-20 13:47:09.981664] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:27:03.099 pt2 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:03.099 [2024-11-20 13:47:09.987364] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:27:03.099 [2024-11-20 13:47:09.989288] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:27:03.099 [2024-11-20 13:47:09.989460] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:27:03.099 [2024-11-20 13:47:09.989471] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:27:03.099 [2024-11-20 13:47:09.989725] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:27:03.099 [2024-11-20 13:47:09.989860] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:27:03.099 [2024-11-20 13:47:09.989870] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:27:03.099 [2024-11-20 13:47:09.990026] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:03.099 13:47:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:03.099 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:03.099 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:03.099 "name": "raid_bdev1", 00:27:03.099 "uuid": "da3b391b-ba1e-4e8e-b066-9aa614a6d3bc", 00:27:03.099 "strip_size_kb": 64, 00:27:03.099 "state": "online", 00:27:03.099 "raid_level": "concat", 00:27:03.099 "superblock": true, 00:27:03.099 "num_base_bdevs": 2, 00:27:03.099 "num_base_bdevs_discovered": 2, 00:27:03.099 "num_base_bdevs_operational": 2, 00:27:03.099 "base_bdevs_list": [ 00:27:03.099 { 00:27:03.099 "name": "pt1", 00:27:03.099 "uuid": "00000000-0000-0000-0000-000000000001", 00:27:03.099 "is_configured": true, 00:27:03.099 "data_offset": 2048, 00:27:03.099 "data_size": 63488 00:27:03.099 }, 00:27:03.099 { 00:27:03.099 "name": "pt2", 00:27:03.099 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:03.099 "is_configured": true, 00:27:03.099 "data_offset": 2048, 00:27:03.099 "data_size": 63488 00:27:03.099 } 00:27:03.099 ] 00:27:03.099 }' 00:27:03.099 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:03.099 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:03.357 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:27:03.357 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:27:03.357 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:27:03.357 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:27:03.357 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:27:03.357 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:27:03.357 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:27:03.357 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:03.357 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:03.357 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:27:03.357 [2024-11-20 13:47:10.311764] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:03.357 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:03.357 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:27:03.357 "name": "raid_bdev1", 00:27:03.357 "aliases": [ 00:27:03.357 "da3b391b-ba1e-4e8e-b066-9aa614a6d3bc" 00:27:03.357 ], 00:27:03.357 "product_name": "Raid Volume", 00:27:03.357 "block_size": 512, 00:27:03.357 "num_blocks": 126976, 00:27:03.357 "uuid": "da3b391b-ba1e-4e8e-b066-9aa614a6d3bc", 00:27:03.357 "assigned_rate_limits": { 00:27:03.357 "rw_ios_per_sec": 0, 00:27:03.357 "rw_mbytes_per_sec": 0, 00:27:03.357 "r_mbytes_per_sec": 0, 00:27:03.357 "w_mbytes_per_sec": 0 00:27:03.357 }, 00:27:03.357 "claimed": false, 00:27:03.357 "zoned": false, 00:27:03.357 "supported_io_types": { 00:27:03.357 "read": true, 00:27:03.357 "write": true, 00:27:03.357 "unmap": true, 00:27:03.357 "flush": true, 00:27:03.357 "reset": true, 00:27:03.357 "nvme_admin": false, 00:27:03.357 "nvme_io": false, 00:27:03.357 "nvme_io_md": false, 00:27:03.357 "write_zeroes": true, 00:27:03.357 "zcopy": false, 00:27:03.357 "get_zone_info": false, 00:27:03.357 "zone_management": false, 00:27:03.357 "zone_append": false, 00:27:03.357 "compare": false, 00:27:03.357 "compare_and_write": false, 00:27:03.357 "abort": false, 00:27:03.357 "seek_hole": false, 00:27:03.357 "seek_data": false, 00:27:03.357 "copy": false, 00:27:03.357 "nvme_iov_md": false 00:27:03.357 }, 00:27:03.357 "memory_domains": [ 00:27:03.357 { 00:27:03.357 "dma_device_id": "system", 00:27:03.357 "dma_device_type": 1 00:27:03.357 }, 00:27:03.357 { 00:27:03.357 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:03.357 "dma_device_type": 2 00:27:03.357 }, 00:27:03.357 { 00:27:03.357 "dma_device_id": "system", 00:27:03.357 "dma_device_type": 1 00:27:03.357 }, 00:27:03.357 { 00:27:03.357 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:03.357 "dma_device_type": 2 00:27:03.358 } 00:27:03.358 ], 00:27:03.358 "driver_specific": { 00:27:03.358 "raid": { 00:27:03.358 "uuid": "da3b391b-ba1e-4e8e-b066-9aa614a6d3bc", 00:27:03.358 "strip_size_kb": 64, 00:27:03.358 "state": "online", 00:27:03.358 "raid_level": "concat", 00:27:03.358 "superblock": true, 00:27:03.358 "num_base_bdevs": 2, 00:27:03.358 "num_base_bdevs_discovered": 2, 00:27:03.358 "num_base_bdevs_operational": 2, 00:27:03.358 "base_bdevs_list": [ 00:27:03.358 { 00:27:03.358 "name": "pt1", 00:27:03.358 "uuid": "00000000-0000-0000-0000-000000000001", 00:27:03.358 "is_configured": true, 00:27:03.358 "data_offset": 2048, 00:27:03.358 "data_size": 63488 00:27:03.358 }, 00:27:03.358 { 00:27:03.358 "name": "pt2", 00:27:03.358 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:03.358 "is_configured": true, 00:27:03.358 "data_offset": 2048, 00:27:03.358 "data_size": 63488 00:27:03.358 } 00:27:03.358 ] 00:27:03.358 } 00:27:03.358 } 00:27:03.358 }' 00:27:03.358 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:27:03.358 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:27:03.358 pt2' 00:27:03.358 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:03.358 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:27:03.358 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:03.358 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:27:03.358 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:03.358 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:03.358 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:03.617 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:03.617 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:03.617 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:03.618 [2024-11-20 13:47:10.483774] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=da3b391b-ba1e-4e8e-b066-9aa614a6d3bc 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z da3b391b-ba1e-4e8e-b066-9aa614a6d3bc ']' 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:03.618 [2024-11-20 13:47:10.507440] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:27:03.618 [2024-11-20 13:47:10.507469] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:27:03.618 [2024-11-20 13:47:10.507560] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:03.618 [2024-11-20 13:47:10.507619] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:03.618 [2024-11-20 13:47:10.507632] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:03.618 [2024-11-20 13:47:10.603534] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:27:03.618 [2024-11-20 13:47:10.605610] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:27:03.618 [2024-11-20 13:47:10.605687] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:27:03.618 [2024-11-20 13:47:10.605743] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:27:03.618 [2024-11-20 13:47:10.605758] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:27:03.618 [2024-11-20 13:47:10.605770] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:27:03.618 request: 00:27:03.618 { 00:27:03.618 "name": "raid_bdev1", 00:27:03.618 "raid_level": "concat", 00:27:03.618 "base_bdevs": [ 00:27:03.618 "malloc1", 00:27:03.618 "malloc2" 00:27:03.618 ], 00:27:03.618 "strip_size_kb": 64, 00:27:03.618 "superblock": false, 00:27:03.618 "method": "bdev_raid_create", 00:27:03.618 "req_id": 1 00:27:03.618 } 00:27:03.618 Got JSON-RPC error response 00:27:03.618 response: 00:27:03.618 { 00:27:03.618 "code": -17, 00:27:03.618 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:27:03.618 } 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:03.618 [2024-11-20 13:47:10.647512] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:27:03.618 [2024-11-20 13:47:10.647577] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:03.618 [2024-11-20 13:47:10.647595] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:27:03.618 [2024-11-20 13:47:10.647606] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:03.618 [2024-11-20 13:47:10.649993] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:03.618 [2024-11-20 13:47:10.650029] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:27:03.618 [2024-11-20 13:47:10.650122] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:27:03.618 [2024-11-20 13:47:10.650178] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:27:03.618 pt1 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 2 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:03.618 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:03.875 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:03.875 "name": "raid_bdev1", 00:27:03.875 "uuid": "da3b391b-ba1e-4e8e-b066-9aa614a6d3bc", 00:27:03.875 "strip_size_kb": 64, 00:27:03.875 "state": "configuring", 00:27:03.875 "raid_level": "concat", 00:27:03.875 "superblock": true, 00:27:03.875 "num_base_bdevs": 2, 00:27:03.875 "num_base_bdevs_discovered": 1, 00:27:03.875 "num_base_bdevs_operational": 2, 00:27:03.875 "base_bdevs_list": [ 00:27:03.875 { 00:27:03.875 "name": "pt1", 00:27:03.875 "uuid": "00000000-0000-0000-0000-000000000001", 00:27:03.875 "is_configured": true, 00:27:03.875 "data_offset": 2048, 00:27:03.875 "data_size": 63488 00:27:03.875 }, 00:27:03.875 { 00:27:03.875 "name": null, 00:27:03.875 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:03.875 "is_configured": false, 00:27:03.875 "data_offset": 2048, 00:27:03.875 "data_size": 63488 00:27:03.875 } 00:27:03.875 ] 00:27:03.875 }' 00:27:03.875 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:03.875 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:04.133 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:27:04.133 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:27:04.133 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:27:04.133 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:27:04.133 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:04.133 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:04.133 [2024-11-20 13:47:10.959604] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:27:04.133 [2024-11-20 13:47:10.959682] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:04.133 [2024-11-20 13:47:10.959703] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:27:04.133 [2024-11-20 13:47:10.959715] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:04.133 [2024-11-20 13:47:10.960199] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:04.133 [2024-11-20 13:47:10.960224] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:27:04.133 [2024-11-20 13:47:10.960304] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:27:04.133 [2024-11-20 13:47:10.960333] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:27:04.133 [2024-11-20 13:47:10.960443] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:27:04.133 [2024-11-20 13:47:10.960455] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:27:04.133 [2024-11-20 13:47:10.960693] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:27:04.133 [2024-11-20 13:47:10.960816] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:27:04.133 [2024-11-20 13:47:10.960824] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:27:04.133 [2024-11-20 13:47:10.960956] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:04.133 pt2 00:27:04.133 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:04.133 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:27:04.133 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:27:04.133 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:27:04.133 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:04.133 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:04.133 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:27:04.133 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:04.133 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:04.133 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:04.133 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:04.133 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:04.133 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:04.133 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:04.133 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:04.133 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:04.133 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:04.133 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:04.133 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:04.133 "name": "raid_bdev1", 00:27:04.133 "uuid": "da3b391b-ba1e-4e8e-b066-9aa614a6d3bc", 00:27:04.134 "strip_size_kb": 64, 00:27:04.134 "state": "online", 00:27:04.134 "raid_level": "concat", 00:27:04.134 "superblock": true, 00:27:04.134 "num_base_bdevs": 2, 00:27:04.134 "num_base_bdevs_discovered": 2, 00:27:04.134 "num_base_bdevs_operational": 2, 00:27:04.134 "base_bdevs_list": [ 00:27:04.134 { 00:27:04.134 "name": "pt1", 00:27:04.134 "uuid": "00000000-0000-0000-0000-000000000001", 00:27:04.134 "is_configured": true, 00:27:04.134 "data_offset": 2048, 00:27:04.134 "data_size": 63488 00:27:04.134 }, 00:27:04.134 { 00:27:04.134 "name": "pt2", 00:27:04.134 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:04.134 "is_configured": true, 00:27:04.134 "data_offset": 2048, 00:27:04.134 "data_size": 63488 00:27:04.134 } 00:27:04.134 ] 00:27:04.134 }' 00:27:04.134 13:47:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:04.134 13:47:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:04.392 [2024-11-20 13:47:11.291947] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:27:04.392 "name": "raid_bdev1", 00:27:04.392 "aliases": [ 00:27:04.392 "da3b391b-ba1e-4e8e-b066-9aa614a6d3bc" 00:27:04.392 ], 00:27:04.392 "product_name": "Raid Volume", 00:27:04.392 "block_size": 512, 00:27:04.392 "num_blocks": 126976, 00:27:04.392 "uuid": "da3b391b-ba1e-4e8e-b066-9aa614a6d3bc", 00:27:04.392 "assigned_rate_limits": { 00:27:04.392 "rw_ios_per_sec": 0, 00:27:04.392 "rw_mbytes_per_sec": 0, 00:27:04.392 "r_mbytes_per_sec": 0, 00:27:04.392 "w_mbytes_per_sec": 0 00:27:04.392 }, 00:27:04.392 "claimed": false, 00:27:04.392 "zoned": false, 00:27:04.392 "supported_io_types": { 00:27:04.392 "read": true, 00:27:04.392 "write": true, 00:27:04.392 "unmap": true, 00:27:04.392 "flush": true, 00:27:04.392 "reset": true, 00:27:04.392 "nvme_admin": false, 00:27:04.392 "nvme_io": false, 00:27:04.392 "nvme_io_md": false, 00:27:04.392 "write_zeroes": true, 00:27:04.392 "zcopy": false, 00:27:04.392 "get_zone_info": false, 00:27:04.392 "zone_management": false, 00:27:04.392 "zone_append": false, 00:27:04.392 "compare": false, 00:27:04.392 "compare_and_write": false, 00:27:04.392 "abort": false, 00:27:04.392 "seek_hole": false, 00:27:04.392 "seek_data": false, 00:27:04.392 "copy": false, 00:27:04.392 "nvme_iov_md": false 00:27:04.392 }, 00:27:04.392 "memory_domains": [ 00:27:04.392 { 00:27:04.392 "dma_device_id": "system", 00:27:04.392 "dma_device_type": 1 00:27:04.392 }, 00:27:04.392 { 00:27:04.392 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:04.392 "dma_device_type": 2 00:27:04.392 }, 00:27:04.392 { 00:27:04.392 "dma_device_id": "system", 00:27:04.392 "dma_device_type": 1 00:27:04.392 }, 00:27:04.392 { 00:27:04.392 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:04.392 "dma_device_type": 2 00:27:04.392 } 00:27:04.392 ], 00:27:04.392 "driver_specific": { 00:27:04.392 "raid": { 00:27:04.392 "uuid": "da3b391b-ba1e-4e8e-b066-9aa614a6d3bc", 00:27:04.392 "strip_size_kb": 64, 00:27:04.392 "state": "online", 00:27:04.392 "raid_level": "concat", 00:27:04.392 "superblock": true, 00:27:04.392 "num_base_bdevs": 2, 00:27:04.392 "num_base_bdevs_discovered": 2, 00:27:04.392 "num_base_bdevs_operational": 2, 00:27:04.392 "base_bdevs_list": [ 00:27:04.392 { 00:27:04.392 "name": "pt1", 00:27:04.392 "uuid": "00000000-0000-0000-0000-000000000001", 00:27:04.392 "is_configured": true, 00:27:04.392 "data_offset": 2048, 00:27:04.392 "data_size": 63488 00:27:04.392 }, 00:27:04.392 { 00:27:04.392 "name": "pt2", 00:27:04.392 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:04.392 "is_configured": true, 00:27:04.392 "data_offset": 2048, 00:27:04.392 "data_size": 63488 00:27:04.392 } 00:27:04.392 ] 00:27:04.392 } 00:27:04.392 } 00:27:04.392 }' 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:27:04.392 pt2' 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:04.392 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:04.650 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:27:04.650 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:27:04.650 13:47:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:04.650 13:47:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:04.650 [2024-11-20 13:47:11.451931] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:04.650 13:47:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:04.650 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' da3b391b-ba1e-4e8e-b066-9aa614a6d3bc '!=' da3b391b-ba1e-4e8e-b066-9aa614a6d3bc ']' 00:27:04.650 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:27:04.650 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:27:04.650 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:27:04.650 13:47:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 60829 00:27:04.650 13:47:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 60829 ']' 00:27:04.650 13:47:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 60829 00:27:04.650 13:47:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:27:04.650 13:47:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:04.650 13:47:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 60829 00:27:04.650 13:47:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:27:04.650 13:47:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:27:04.650 killing process with pid 60829 00:27:04.650 13:47:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 60829' 00:27:04.650 13:47:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 60829 00:27:04.650 [2024-11-20 13:47:11.501205] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:27:04.650 13:47:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 60829 00:27:04.650 [2024-11-20 13:47:11.501292] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:04.650 [2024-11-20 13:47:11.501348] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:04.650 [2024-11-20 13:47:11.501364] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:27:04.650 [2024-11-20 13:47:11.637488] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:27:05.580 13:47:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:27:05.580 00:27:05.580 real 0m3.343s 00:27:05.580 user 0m4.675s 00:27:05.580 sys 0m0.563s 00:27:05.580 13:47:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:27:05.580 13:47:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:05.580 ************************************ 00:27:05.580 END TEST raid_superblock_test 00:27:05.580 ************************************ 00:27:05.580 13:47:12 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 2 read 00:27:05.580 13:47:12 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:27:05.580 13:47:12 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:27:05.580 13:47:12 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:27:05.580 ************************************ 00:27:05.580 START TEST raid_read_error_test 00:27:05.580 ************************************ 00:27:05.580 13:47:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 2 read 00:27:05.580 13:47:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:27:05.580 13:47:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:27:05.580 13:47:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:27:05.580 13:47:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:27:05.580 13:47:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:27:05.580 13:47:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:27:05.580 13:47:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:27:05.580 13:47:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:27:05.580 13:47:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:27:05.580 13:47:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:27:05.580 13:47:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:27:05.580 13:47:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:27:05.580 13:47:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:27:05.580 13:47:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:27:05.580 13:47:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:27:05.580 13:47:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:27:05.580 13:47:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:27:05.580 13:47:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:27:05.581 13:47:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:27:05.581 13:47:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:27:05.581 13:47:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:27:05.581 13:47:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:27:05.581 13:47:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.FPNxpgDrhC 00:27:05.581 13:47:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=61030 00:27:05.581 13:47:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 61030 00:27:05.581 13:47:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 61030 ']' 00:27:05.581 13:47:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:05.581 13:47:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:05.581 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:05.581 13:47:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:05.581 13:47:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:05.581 13:47:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:05.581 13:47:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:27:05.581 [2024-11-20 13:47:12.415229] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:27:05.581 [2024-11-20 13:47:12.415369] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid61030 ] 00:27:05.581 [2024-11-20 13:47:12.573299] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:05.882 [2024-11-20 13:47:12.681707] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:27:05.882 [2024-11-20 13:47:12.803069] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:05.882 [2024-11-20 13:47:12.803128] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:06.471 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:06.471 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:27:06.471 13:47:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:27:06.471 13:47:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:27:06.471 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:06.471 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:06.471 BaseBdev1_malloc 00:27:06.471 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:06.471 13:47:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:27:06.471 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:06.471 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:06.471 true 00:27:06.471 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:06.472 13:47:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:27:06.472 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:06.472 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:06.472 [2024-11-20 13:47:13.298137] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:27:06.472 [2024-11-20 13:47:13.298195] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:06.472 [2024-11-20 13:47:13.298213] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:27:06.472 [2024-11-20 13:47:13.298223] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:06.472 [2024-11-20 13:47:13.300123] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:06.472 [2024-11-20 13:47:13.300157] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:27:06.472 BaseBdev1 00:27:06.472 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:06.472 13:47:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:27:06.472 13:47:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:27:06.472 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:06.472 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:06.472 BaseBdev2_malloc 00:27:06.472 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:06.472 13:47:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:27:06.472 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:06.472 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:06.472 true 00:27:06.472 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:06.472 13:47:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:27:06.472 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:06.472 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:06.472 [2024-11-20 13:47:13.339967] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:27:06.472 [2024-11-20 13:47:13.340030] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:06.472 [2024-11-20 13:47:13.340045] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:27:06.472 [2024-11-20 13:47:13.340053] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:06.472 [2024-11-20 13:47:13.341895] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:06.472 [2024-11-20 13:47:13.341929] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:27:06.472 BaseBdev2 00:27:06.472 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:06.472 13:47:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:27:06.472 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:06.472 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:06.472 [2024-11-20 13:47:13.348048] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:27:06.472 [2024-11-20 13:47:13.349666] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:27:06.472 [2024-11-20 13:47:13.349823] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:27:06.472 [2024-11-20 13:47:13.349835] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:27:06.472 [2024-11-20 13:47:13.350052] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:27:06.472 [2024-11-20 13:47:13.350192] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:27:06.472 [2024-11-20 13:47:13.350201] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:27:06.472 [2024-11-20 13:47:13.350315] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:06.472 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:06.472 13:47:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:27:06.473 13:47:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:06.473 13:47:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:06.473 13:47:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:27:06.473 13:47:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:06.473 13:47:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:06.473 13:47:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:06.473 13:47:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:06.473 13:47:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:06.473 13:47:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:06.473 13:47:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:06.473 13:47:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:06.473 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:06.473 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:06.473 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:06.473 13:47:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:06.473 "name": "raid_bdev1", 00:27:06.473 "uuid": "e2517005-bd63-4ada-bd93-3c8080787a00", 00:27:06.473 "strip_size_kb": 64, 00:27:06.473 "state": "online", 00:27:06.473 "raid_level": "concat", 00:27:06.473 "superblock": true, 00:27:06.473 "num_base_bdevs": 2, 00:27:06.473 "num_base_bdevs_discovered": 2, 00:27:06.473 "num_base_bdevs_operational": 2, 00:27:06.473 "base_bdevs_list": [ 00:27:06.473 { 00:27:06.473 "name": "BaseBdev1", 00:27:06.473 "uuid": "eab3e289-05da-5b4c-90b8-29daedca2052", 00:27:06.473 "is_configured": true, 00:27:06.473 "data_offset": 2048, 00:27:06.473 "data_size": 63488 00:27:06.473 }, 00:27:06.473 { 00:27:06.473 "name": "BaseBdev2", 00:27:06.473 "uuid": "b2df0b95-7d30-5107-86aa-2c43047b85b3", 00:27:06.473 "is_configured": true, 00:27:06.473 "data_offset": 2048, 00:27:06.473 "data_size": 63488 00:27:06.473 } 00:27:06.473 ] 00:27:06.473 }' 00:27:06.473 13:47:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:06.473 13:47:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:06.732 13:47:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:27:06.732 13:47:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:27:06.732 [2024-11-20 13:47:13.744999] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:27:07.667 13:47:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:27:07.667 13:47:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:07.667 13:47:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:07.667 13:47:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:07.667 13:47:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:27:07.667 13:47:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:27:07.667 13:47:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:27:07.667 13:47:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:27:07.667 13:47:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:07.667 13:47:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:07.667 13:47:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:27:07.667 13:47:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:07.667 13:47:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:07.667 13:47:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:07.667 13:47:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:07.667 13:47:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:07.667 13:47:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:07.667 13:47:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:07.667 13:47:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:07.667 13:47:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:07.667 13:47:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:07.667 13:47:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:07.667 13:47:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:07.667 "name": "raid_bdev1", 00:27:07.667 "uuid": "e2517005-bd63-4ada-bd93-3c8080787a00", 00:27:07.667 "strip_size_kb": 64, 00:27:07.667 "state": "online", 00:27:07.667 "raid_level": "concat", 00:27:07.667 "superblock": true, 00:27:07.667 "num_base_bdevs": 2, 00:27:07.667 "num_base_bdevs_discovered": 2, 00:27:07.667 "num_base_bdevs_operational": 2, 00:27:07.667 "base_bdevs_list": [ 00:27:07.667 { 00:27:07.667 "name": "BaseBdev1", 00:27:07.667 "uuid": "eab3e289-05da-5b4c-90b8-29daedca2052", 00:27:07.667 "is_configured": true, 00:27:07.667 "data_offset": 2048, 00:27:07.667 "data_size": 63488 00:27:07.667 }, 00:27:07.667 { 00:27:07.667 "name": "BaseBdev2", 00:27:07.667 "uuid": "b2df0b95-7d30-5107-86aa-2c43047b85b3", 00:27:07.667 "is_configured": true, 00:27:07.667 "data_offset": 2048, 00:27:07.667 "data_size": 63488 00:27:07.667 } 00:27:07.668 ] 00:27:07.668 }' 00:27:07.668 13:47:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:07.668 13:47:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:08.233 13:47:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:27:08.233 13:47:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:08.233 13:47:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:08.233 [2024-11-20 13:47:15.009074] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:27:08.233 [2024-11-20 13:47:15.009116] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:27:08.233 [2024-11-20 13:47:15.011677] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:08.233 [2024-11-20 13:47:15.011726] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:08.233 [2024-11-20 13:47:15.011757] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:08.233 [2024-11-20 13:47:15.011770] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:27:08.233 { 00:27:08.233 "results": [ 00:27:08.233 { 00:27:08.233 "job": "raid_bdev1", 00:27:08.233 "core_mask": "0x1", 00:27:08.233 "workload": "randrw", 00:27:08.233 "percentage": 50, 00:27:08.233 "status": "finished", 00:27:08.233 "queue_depth": 1, 00:27:08.233 "io_size": 131072, 00:27:08.233 "runtime": 1.262485, 00:27:08.233 "iops": 16688.51511107063, 00:27:08.233 "mibps": 2086.0643888838285, 00:27:08.233 "io_failed": 1, 00:27:08.233 "io_timeout": 0, 00:27:08.233 "avg_latency_us": 82.2506102004308, 00:27:08.233 "min_latency_us": 26.38769230769231, 00:27:08.233 "max_latency_us": 1468.2584615384615 00:27:08.233 } 00:27:08.233 ], 00:27:08.233 "core_count": 1 00:27:08.233 } 00:27:08.233 13:47:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:08.233 13:47:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 61030 00:27:08.233 13:47:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 61030 ']' 00:27:08.233 13:47:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 61030 00:27:08.233 13:47:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:27:08.233 13:47:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:08.233 13:47:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 61030 00:27:08.233 13:47:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:27:08.234 13:47:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:27:08.234 killing process with pid 61030 00:27:08.234 13:47:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 61030' 00:27:08.234 13:47:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 61030 00:27:08.234 [2024-11-20 13:47:15.040291] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:27:08.234 13:47:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 61030 00:27:08.234 [2024-11-20 13:47:15.111696] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:27:08.799 13:47:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.FPNxpgDrhC 00:27:08.799 13:47:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:27:08.799 13:47:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:27:08.799 13:47:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.79 00:27:08.799 13:47:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:27:08.799 13:47:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:27:08.799 13:47:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:27:08.799 13:47:15 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.79 != \0\.\0\0 ]] 00:27:08.799 00:27:08.799 real 0m3.424s 00:27:08.799 user 0m4.099s 00:27:08.799 sys 0m0.411s 00:27:08.799 13:47:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:27:08.799 13:47:15 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:08.799 ************************************ 00:27:08.799 END TEST raid_read_error_test 00:27:08.799 ************************************ 00:27:08.799 13:47:15 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 2 write 00:27:08.799 13:47:15 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:27:08.799 13:47:15 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:27:08.799 13:47:15 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:27:08.799 ************************************ 00:27:08.799 START TEST raid_write_error_test 00:27:08.799 ************************************ 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 2 write 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.AfzBMVexF6 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=61159 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 61159 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 61159 ']' 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:08.799 13:47:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:08.799 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:08.800 13:47:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:08.800 13:47:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:08.800 13:47:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:09.057 [2024-11-20 13:47:15.872240] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:27:09.057 [2024-11-20 13:47:15.872359] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid61159 ] 00:27:09.057 [2024-11-20 13:47:16.027384] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:09.315 [2024-11-20 13:47:16.143694] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:27:09.315 [2024-11-20 13:47:16.298605] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:09.315 [2024-11-20 13:47:16.298657] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:09.880 BaseBdev1_malloc 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:09.880 true 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:09.880 [2024-11-20 13:47:16.781192] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:27:09.880 [2024-11-20 13:47:16.781251] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:09.880 [2024-11-20 13:47:16.781270] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:27:09.880 [2024-11-20 13:47:16.781281] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:09.880 [2024-11-20 13:47:16.783511] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:09.880 [2024-11-20 13:47:16.783549] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:27:09.880 BaseBdev1 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:09.880 BaseBdev2_malloc 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:09.880 true 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:09.880 [2024-11-20 13:47:16.827191] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:27:09.880 [2024-11-20 13:47:16.827247] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:09.880 [2024-11-20 13:47:16.827263] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:27:09.880 [2024-11-20 13:47:16.827274] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:09.880 [2024-11-20 13:47:16.829506] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:09.880 [2024-11-20 13:47:16.829544] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:27:09.880 BaseBdev2 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:09.880 [2024-11-20 13:47:16.835268] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:27:09.880 [2024-11-20 13:47:16.837209] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:27:09.880 [2024-11-20 13:47:16.837401] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:27:09.880 [2024-11-20 13:47:16.837415] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:27:09.880 [2024-11-20 13:47:16.837665] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:27:09.880 [2024-11-20 13:47:16.837828] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:27:09.880 [2024-11-20 13:47:16.837840] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:27:09.880 [2024-11-20 13:47:16.837998] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:09.880 13:47:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:09.881 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:09.881 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:09.881 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:09.881 13:47:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:09.881 "name": "raid_bdev1", 00:27:09.881 "uuid": "ed0e5a83-7f9a-49fc-8133-bffb7e8c8f29", 00:27:09.881 "strip_size_kb": 64, 00:27:09.881 "state": "online", 00:27:09.881 "raid_level": "concat", 00:27:09.881 "superblock": true, 00:27:09.881 "num_base_bdevs": 2, 00:27:09.881 "num_base_bdevs_discovered": 2, 00:27:09.881 "num_base_bdevs_operational": 2, 00:27:09.881 "base_bdevs_list": [ 00:27:09.881 { 00:27:09.881 "name": "BaseBdev1", 00:27:09.881 "uuid": "211fba34-8a44-54dd-b441-91a86fcb0bda", 00:27:09.881 "is_configured": true, 00:27:09.881 "data_offset": 2048, 00:27:09.881 "data_size": 63488 00:27:09.881 }, 00:27:09.881 { 00:27:09.881 "name": "BaseBdev2", 00:27:09.881 "uuid": "5e01b430-af70-53c9-9e9d-a44defb934f5", 00:27:09.881 "is_configured": true, 00:27:09.881 "data_offset": 2048, 00:27:09.881 "data_size": 63488 00:27:09.881 } 00:27:09.881 ] 00:27:09.881 }' 00:27:09.881 13:47:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:09.881 13:47:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:10.167 13:47:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:27:10.167 13:47:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:27:10.425 [2024-11-20 13:47:17.224424] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:27:11.358 13:47:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:27:11.358 13:47:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:11.358 13:47:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:11.359 13:47:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:11.359 13:47:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:27:11.359 13:47:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:27:11.359 13:47:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:27:11.359 13:47:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:27:11.359 13:47:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:11.359 13:47:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:11.359 13:47:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:27:11.359 13:47:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:11.359 13:47:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:11.359 13:47:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:11.359 13:47:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:11.359 13:47:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:11.359 13:47:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:11.359 13:47:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:11.359 13:47:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:11.359 13:47:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:11.359 13:47:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:11.359 13:47:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:11.359 13:47:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:11.359 "name": "raid_bdev1", 00:27:11.359 "uuid": "ed0e5a83-7f9a-49fc-8133-bffb7e8c8f29", 00:27:11.359 "strip_size_kb": 64, 00:27:11.359 "state": "online", 00:27:11.359 "raid_level": "concat", 00:27:11.359 "superblock": true, 00:27:11.359 "num_base_bdevs": 2, 00:27:11.359 "num_base_bdevs_discovered": 2, 00:27:11.359 "num_base_bdevs_operational": 2, 00:27:11.359 "base_bdevs_list": [ 00:27:11.359 { 00:27:11.359 "name": "BaseBdev1", 00:27:11.359 "uuid": "211fba34-8a44-54dd-b441-91a86fcb0bda", 00:27:11.359 "is_configured": true, 00:27:11.359 "data_offset": 2048, 00:27:11.359 "data_size": 63488 00:27:11.359 }, 00:27:11.359 { 00:27:11.359 "name": "BaseBdev2", 00:27:11.359 "uuid": "5e01b430-af70-53c9-9e9d-a44defb934f5", 00:27:11.359 "is_configured": true, 00:27:11.359 "data_offset": 2048, 00:27:11.359 "data_size": 63488 00:27:11.359 } 00:27:11.359 ] 00:27:11.359 }' 00:27:11.359 13:47:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:11.359 13:47:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:11.618 13:47:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:27:11.618 13:47:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:11.618 13:47:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:11.618 [2024-11-20 13:47:18.458723] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:27:11.618 [2024-11-20 13:47:18.458762] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:27:11.618 [2024-11-20 13:47:18.461913] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:11.618 [2024-11-20 13:47:18.461963] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:11.618 [2024-11-20 13:47:18.462007] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:11.618 [2024-11-20 13:47:18.462019] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:27:11.618 13:47:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:11.618 { 00:27:11.618 "results": [ 00:27:11.618 { 00:27:11.618 "job": "raid_bdev1", 00:27:11.618 "core_mask": "0x1", 00:27:11.618 "workload": "randrw", 00:27:11.618 "percentage": 50, 00:27:11.618 "status": "finished", 00:27:11.618 "queue_depth": 1, 00:27:11.618 "io_size": 131072, 00:27:11.618 "runtime": 1.232443, 00:27:11.618 "iops": 14040.405925466735, 00:27:11.618 "mibps": 1755.050740683342, 00:27:11.618 "io_failed": 1, 00:27:11.618 "io_timeout": 0, 00:27:11.618 "avg_latency_us": 97.29268250616762, 00:27:11.618 "min_latency_us": 33.673846153846156, 00:27:11.618 "max_latency_us": 1739.2246153846154 00:27:11.618 } 00:27:11.618 ], 00:27:11.618 "core_count": 1 00:27:11.618 } 00:27:11.618 13:47:18 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 61159 00:27:11.618 13:47:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 61159 ']' 00:27:11.618 13:47:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 61159 00:27:11.618 13:47:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:27:11.618 13:47:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:11.618 13:47:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 61159 00:27:11.618 killing process with pid 61159 00:27:11.618 13:47:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:27:11.618 13:47:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:27:11.618 13:47:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 61159' 00:27:11.618 13:47:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 61159 00:27:11.618 [2024-11-20 13:47:18.492237] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:27:11.618 13:47:18 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 61159 00:27:11.618 [2024-11-20 13:47:18.581608] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:27:12.551 13:47:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.AfzBMVexF6 00:27:12.551 13:47:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:27:12.551 13:47:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:27:12.551 ************************************ 00:27:12.551 END TEST raid_write_error_test 00:27:12.551 ************************************ 00:27:12.551 13:47:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.81 00:27:12.551 13:47:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:27:12.551 13:47:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:27:12.551 13:47:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:27:12.551 13:47:19 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.81 != \0\.\0\0 ]] 00:27:12.551 00:27:12.551 real 0m3.472s 00:27:12.551 user 0m4.138s 00:27:12.551 sys 0m0.409s 00:27:12.551 13:47:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:27:12.551 13:47:19 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:12.551 13:47:19 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:27:12.551 13:47:19 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 2 false 00:27:12.551 13:47:19 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:27:12.551 13:47:19 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:27:12.551 13:47:19 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:27:12.551 ************************************ 00:27:12.551 START TEST raid_state_function_test 00:27:12.551 ************************************ 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 false 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:27:12.551 Process raid pid: 61291 00:27:12.551 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=61291 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 61291' 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 61291 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 61291 ']' 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:12.551 13:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:27:12.551 [2024-11-20 13:47:19.386392] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:27:12.551 [2024-11-20 13:47:19.386514] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:27:12.551 [2024-11-20 13:47:19.541708] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:12.810 [2024-11-20 13:47:19.644461] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:27:12.810 [2024-11-20 13:47:19.767496] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:12.810 [2024-11-20 13:47:19.767707] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:13.377 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:13.377 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:27:13.377 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:27:13.377 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:13.377 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:13.377 [2024-11-20 13:47:20.226160] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:27:13.377 [2024-11-20 13:47:20.226220] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:27:13.377 [2024-11-20 13:47:20.226229] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:27:13.378 [2024-11-20 13:47:20.226238] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:27:13.378 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:13.378 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:27:13.378 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:13.378 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:13.378 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:13.378 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:13.378 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:13.378 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:13.378 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:13.378 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:13.378 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:13.378 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:13.378 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:13.378 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:13.378 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:13.378 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:13.378 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:13.378 "name": "Existed_Raid", 00:27:13.378 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:13.378 "strip_size_kb": 0, 00:27:13.378 "state": "configuring", 00:27:13.378 "raid_level": "raid1", 00:27:13.378 "superblock": false, 00:27:13.378 "num_base_bdevs": 2, 00:27:13.378 "num_base_bdevs_discovered": 0, 00:27:13.378 "num_base_bdevs_operational": 2, 00:27:13.378 "base_bdevs_list": [ 00:27:13.378 { 00:27:13.378 "name": "BaseBdev1", 00:27:13.378 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:13.378 "is_configured": false, 00:27:13.378 "data_offset": 0, 00:27:13.378 "data_size": 0 00:27:13.378 }, 00:27:13.378 { 00:27:13.378 "name": "BaseBdev2", 00:27:13.378 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:13.378 "is_configured": false, 00:27:13.378 "data_offset": 0, 00:27:13.378 "data_size": 0 00:27:13.378 } 00:27:13.378 ] 00:27:13.378 }' 00:27:13.378 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:13.378 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:13.637 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:27:13.637 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:13.637 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:13.637 [2024-11-20 13:47:20.546168] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:27:13.637 [2024-11-20 13:47:20.546199] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:27:13.637 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:13.637 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:27:13.637 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:13.637 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:13.637 [2024-11-20 13:47:20.554141] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:27:13.637 [2024-11-20 13:47:20.554176] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:27:13.637 [2024-11-20 13:47:20.554184] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:27:13.637 [2024-11-20 13:47:20.554195] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:27:13.637 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:13.637 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:27:13.637 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:13.637 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:13.637 [2024-11-20 13:47:20.584363] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:27:13.637 BaseBdev1 00:27:13.637 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:13.637 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:13.638 [ 00:27:13.638 { 00:27:13.638 "name": "BaseBdev1", 00:27:13.638 "aliases": [ 00:27:13.638 "02a59260-97db-4e88-933e-d142fed23217" 00:27:13.638 ], 00:27:13.638 "product_name": "Malloc disk", 00:27:13.638 "block_size": 512, 00:27:13.638 "num_blocks": 65536, 00:27:13.638 "uuid": "02a59260-97db-4e88-933e-d142fed23217", 00:27:13.638 "assigned_rate_limits": { 00:27:13.638 "rw_ios_per_sec": 0, 00:27:13.638 "rw_mbytes_per_sec": 0, 00:27:13.638 "r_mbytes_per_sec": 0, 00:27:13.638 "w_mbytes_per_sec": 0 00:27:13.638 }, 00:27:13.638 "claimed": true, 00:27:13.638 "claim_type": "exclusive_write", 00:27:13.638 "zoned": false, 00:27:13.638 "supported_io_types": { 00:27:13.638 "read": true, 00:27:13.638 "write": true, 00:27:13.638 "unmap": true, 00:27:13.638 "flush": true, 00:27:13.638 "reset": true, 00:27:13.638 "nvme_admin": false, 00:27:13.638 "nvme_io": false, 00:27:13.638 "nvme_io_md": false, 00:27:13.638 "write_zeroes": true, 00:27:13.638 "zcopy": true, 00:27:13.638 "get_zone_info": false, 00:27:13.638 "zone_management": false, 00:27:13.638 "zone_append": false, 00:27:13.638 "compare": false, 00:27:13.638 "compare_and_write": false, 00:27:13.638 "abort": true, 00:27:13.638 "seek_hole": false, 00:27:13.638 "seek_data": false, 00:27:13.638 "copy": true, 00:27:13.638 "nvme_iov_md": false 00:27:13.638 }, 00:27:13.638 "memory_domains": [ 00:27:13.638 { 00:27:13.638 "dma_device_id": "system", 00:27:13.638 "dma_device_type": 1 00:27:13.638 }, 00:27:13.638 { 00:27:13.638 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:13.638 "dma_device_type": 2 00:27:13.638 } 00:27:13.638 ], 00:27:13.638 "driver_specific": {} 00:27:13.638 } 00:27:13.638 ] 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:13.638 "name": "Existed_Raid", 00:27:13.638 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:13.638 "strip_size_kb": 0, 00:27:13.638 "state": "configuring", 00:27:13.638 "raid_level": "raid1", 00:27:13.638 "superblock": false, 00:27:13.638 "num_base_bdevs": 2, 00:27:13.638 "num_base_bdevs_discovered": 1, 00:27:13.638 "num_base_bdevs_operational": 2, 00:27:13.638 "base_bdevs_list": [ 00:27:13.638 { 00:27:13.638 "name": "BaseBdev1", 00:27:13.638 "uuid": "02a59260-97db-4e88-933e-d142fed23217", 00:27:13.638 "is_configured": true, 00:27:13.638 "data_offset": 0, 00:27:13.638 "data_size": 65536 00:27:13.638 }, 00:27:13.638 { 00:27:13.638 "name": "BaseBdev2", 00:27:13.638 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:13.638 "is_configured": false, 00:27:13.638 "data_offset": 0, 00:27:13.638 "data_size": 0 00:27:13.638 } 00:27:13.638 ] 00:27:13.638 }' 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:13.638 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:13.897 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:27:13.897 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:13.897 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:13.897 [2024-11-20 13:47:20.900489] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:27:13.897 [2024-11-20 13:47:20.900545] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:27:13.897 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:13.897 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:27:13.897 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:13.897 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:13.897 [2024-11-20 13:47:20.908513] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:27:13.897 [2024-11-20 13:47:20.910198] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:27:13.897 [2024-11-20 13:47:20.910237] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:27:13.897 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:13.897 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:27:13.897 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:27:13.898 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:27:13.898 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:13.898 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:13.898 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:13.898 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:13.898 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:13.898 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:13.898 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:13.898 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:13.898 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:13.898 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:13.898 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:13.898 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:13.898 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:13.898 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:13.898 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:13.898 "name": "Existed_Raid", 00:27:13.898 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:13.898 "strip_size_kb": 0, 00:27:13.898 "state": "configuring", 00:27:13.898 "raid_level": "raid1", 00:27:13.898 "superblock": false, 00:27:13.898 "num_base_bdevs": 2, 00:27:13.898 "num_base_bdevs_discovered": 1, 00:27:13.898 "num_base_bdevs_operational": 2, 00:27:13.898 "base_bdevs_list": [ 00:27:13.898 { 00:27:13.898 "name": "BaseBdev1", 00:27:13.898 "uuid": "02a59260-97db-4e88-933e-d142fed23217", 00:27:13.898 "is_configured": true, 00:27:13.898 "data_offset": 0, 00:27:13.898 "data_size": 65536 00:27:13.898 }, 00:27:13.898 { 00:27:13.898 "name": "BaseBdev2", 00:27:13.898 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:13.898 "is_configured": false, 00:27:13.898 "data_offset": 0, 00:27:13.898 "data_size": 0 00:27:13.898 } 00:27:13.898 ] 00:27:13.898 }' 00:27:13.898 13:47:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:13.898 13:47:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:14.491 [2024-11-20 13:47:21.258129] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:27:14.491 [2024-11-20 13:47:21.258184] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:27:14.491 [2024-11-20 13:47:21.258192] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:27:14.491 [2024-11-20 13:47:21.258420] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:27:14.491 [2024-11-20 13:47:21.258576] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:27:14.491 [2024-11-20 13:47:21.258587] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:27:14.491 [2024-11-20 13:47:21.258813] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:14.491 BaseBdev2 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:14.491 [ 00:27:14.491 { 00:27:14.491 "name": "BaseBdev2", 00:27:14.491 "aliases": [ 00:27:14.491 "85795f5d-1140-469e-ac56-937aca979377" 00:27:14.491 ], 00:27:14.491 "product_name": "Malloc disk", 00:27:14.491 "block_size": 512, 00:27:14.491 "num_blocks": 65536, 00:27:14.491 "uuid": "85795f5d-1140-469e-ac56-937aca979377", 00:27:14.491 "assigned_rate_limits": { 00:27:14.491 "rw_ios_per_sec": 0, 00:27:14.491 "rw_mbytes_per_sec": 0, 00:27:14.491 "r_mbytes_per_sec": 0, 00:27:14.491 "w_mbytes_per_sec": 0 00:27:14.491 }, 00:27:14.491 "claimed": true, 00:27:14.491 "claim_type": "exclusive_write", 00:27:14.491 "zoned": false, 00:27:14.491 "supported_io_types": { 00:27:14.491 "read": true, 00:27:14.491 "write": true, 00:27:14.491 "unmap": true, 00:27:14.491 "flush": true, 00:27:14.491 "reset": true, 00:27:14.491 "nvme_admin": false, 00:27:14.491 "nvme_io": false, 00:27:14.491 "nvme_io_md": false, 00:27:14.491 "write_zeroes": true, 00:27:14.491 "zcopy": true, 00:27:14.491 "get_zone_info": false, 00:27:14.491 "zone_management": false, 00:27:14.491 "zone_append": false, 00:27:14.491 "compare": false, 00:27:14.491 "compare_and_write": false, 00:27:14.491 "abort": true, 00:27:14.491 "seek_hole": false, 00:27:14.491 "seek_data": false, 00:27:14.491 "copy": true, 00:27:14.491 "nvme_iov_md": false 00:27:14.491 }, 00:27:14.491 "memory_domains": [ 00:27:14.491 { 00:27:14.491 "dma_device_id": "system", 00:27:14.491 "dma_device_type": 1 00:27:14.491 }, 00:27:14.491 { 00:27:14.491 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:14.491 "dma_device_type": 2 00:27:14.491 } 00:27:14.491 ], 00:27:14.491 "driver_specific": {} 00:27:14.491 } 00:27:14.491 ] 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:14.491 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:14.491 "name": "Existed_Raid", 00:27:14.491 "uuid": "5895dfa5-3aa0-4680-8374-2be84e1f2887", 00:27:14.491 "strip_size_kb": 0, 00:27:14.491 "state": "online", 00:27:14.491 "raid_level": "raid1", 00:27:14.491 "superblock": false, 00:27:14.491 "num_base_bdevs": 2, 00:27:14.491 "num_base_bdevs_discovered": 2, 00:27:14.491 "num_base_bdevs_operational": 2, 00:27:14.491 "base_bdevs_list": [ 00:27:14.492 { 00:27:14.492 "name": "BaseBdev1", 00:27:14.492 "uuid": "02a59260-97db-4e88-933e-d142fed23217", 00:27:14.492 "is_configured": true, 00:27:14.492 "data_offset": 0, 00:27:14.492 "data_size": 65536 00:27:14.492 }, 00:27:14.492 { 00:27:14.492 "name": "BaseBdev2", 00:27:14.492 "uuid": "85795f5d-1140-469e-ac56-937aca979377", 00:27:14.492 "is_configured": true, 00:27:14.492 "data_offset": 0, 00:27:14.492 "data_size": 65536 00:27:14.492 } 00:27:14.492 ] 00:27:14.492 }' 00:27:14.492 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:14.492 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:14.750 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:27:14.750 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:27:14.750 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:27:14.750 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:27:14.750 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:27:14.750 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:27:14.750 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:27:14.750 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:27:14.750 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:14.750 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:14.750 [2024-11-20 13:47:21.606499] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:14.750 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:14.750 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:27:14.750 "name": "Existed_Raid", 00:27:14.750 "aliases": [ 00:27:14.750 "5895dfa5-3aa0-4680-8374-2be84e1f2887" 00:27:14.750 ], 00:27:14.750 "product_name": "Raid Volume", 00:27:14.750 "block_size": 512, 00:27:14.750 "num_blocks": 65536, 00:27:14.750 "uuid": "5895dfa5-3aa0-4680-8374-2be84e1f2887", 00:27:14.750 "assigned_rate_limits": { 00:27:14.750 "rw_ios_per_sec": 0, 00:27:14.750 "rw_mbytes_per_sec": 0, 00:27:14.750 "r_mbytes_per_sec": 0, 00:27:14.750 "w_mbytes_per_sec": 0 00:27:14.750 }, 00:27:14.750 "claimed": false, 00:27:14.750 "zoned": false, 00:27:14.750 "supported_io_types": { 00:27:14.750 "read": true, 00:27:14.750 "write": true, 00:27:14.750 "unmap": false, 00:27:14.750 "flush": false, 00:27:14.750 "reset": true, 00:27:14.750 "nvme_admin": false, 00:27:14.750 "nvme_io": false, 00:27:14.750 "nvme_io_md": false, 00:27:14.750 "write_zeroes": true, 00:27:14.750 "zcopy": false, 00:27:14.750 "get_zone_info": false, 00:27:14.750 "zone_management": false, 00:27:14.750 "zone_append": false, 00:27:14.750 "compare": false, 00:27:14.750 "compare_and_write": false, 00:27:14.750 "abort": false, 00:27:14.750 "seek_hole": false, 00:27:14.750 "seek_data": false, 00:27:14.750 "copy": false, 00:27:14.750 "nvme_iov_md": false 00:27:14.750 }, 00:27:14.750 "memory_domains": [ 00:27:14.750 { 00:27:14.750 "dma_device_id": "system", 00:27:14.750 "dma_device_type": 1 00:27:14.750 }, 00:27:14.750 { 00:27:14.750 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:14.750 "dma_device_type": 2 00:27:14.750 }, 00:27:14.750 { 00:27:14.750 "dma_device_id": "system", 00:27:14.750 "dma_device_type": 1 00:27:14.750 }, 00:27:14.750 { 00:27:14.750 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:14.750 "dma_device_type": 2 00:27:14.750 } 00:27:14.750 ], 00:27:14.750 "driver_specific": { 00:27:14.750 "raid": { 00:27:14.750 "uuid": "5895dfa5-3aa0-4680-8374-2be84e1f2887", 00:27:14.750 "strip_size_kb": 0, 00:27:14.750 "state": "online", 00:27:14.750 "raid_level": "raid1", 00:27:14.750 "superblock": false, 00:27:14.750 "num_base_bdevs": 2, 00:27:14.750 "num_base_bdevs_discovered": 2, 00:27:14.750 "num_base_bdevs_operational": 2, 00:27:14.750 "base_bdevs_list": [ 00:27:14.750 { 00:27:14.750 "name": "BaseBdev1", 00:27:14.750 "uuid": "02a59260-97db-4e88-933e-d142fed23217", 00:27:14.750 "is_configured": true, 00:27:14.750 "data_offset": 0, 00:27:14.750 "data_size": 65536 00:27:14.750 }, 00:27:14.750 { 00:27:14.750 "name": "BaseBdev2", 00:27:14.750 "uuid": "85795f5d-1140-469e-ac56-937aca979377", 00:27:14.750 "is_configured": true, 00:27:14.750 "data_offset": 0, 00:27:14.750 "data_size": 65536 00:27:14.750 } 00:27:14.750 ] 00:27:14.750 } 00:27:14.750 } 00:27:14.750 }' 00:27:14.750 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:27:14.750 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:27:14.750 BaseBdev2' 00:27:14.751 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:14.751 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:27:14.751 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:14.751 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:27:14.751 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:14.751 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:14.751 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:14.751 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:14.751 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:14.751 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:14.751 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:14.751 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:14.751 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:27:14.751 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:14.751 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:14.751 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:14.751 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:14.751 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:14.751 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:27:14.751 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:14.751 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:14.751 [2024-11-20 13:47:21.798334] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:27:15.008 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:15.008 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:27:15.008 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:27:15.008 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:27:15.008 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:27:15.008 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:27:15.008 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:27:15.008 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:15.008 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:15.008 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:15.008 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:15.008 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:27:15.008 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:15.008 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:15.008 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:15.008 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:15.008 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:15.009 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:15.009 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:15.009 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:15.009 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:15.009 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:15.009 "name": "Existed_Raid", 00:27:15.009 "uuid": "5895dfa5-3aa0-4680-8374-2be84e1f2887", 00:27:15.009 "strip_size_kb": 0, 00:27:15.009 "state": "online", 00:27:15.009 "raid_level": "raid1", 00:27:15.009 "superblock": false, 00:27:15.009 "num_base_bdevs": 2, 00:27:15.009 "num_base_bdevs_discovered": 1, 00:27:15.009 "num_base_bdevs_operational": 1, 00:27:15.009 "base_bdevs_list": [ 00:27:15.009 { 00:27:15.009 "name": null, 00:27:15.009 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:15.009 "is_configured": false, 00:27:15.009 "data_offset": 0, 00:27:15.009 "data_size": 65536 00:27:15.009 }, 00:27:15.009 { 00:27:15.009 "name": "BaseBdev2", 00:27:15.009 "uuid": "85795f5d-1140-469e-ac56-937aca979377", 00:27:15.009 "is_configured": true, 00:27:15.009 "data_offset": 0, 00:27:15.009 "data_size": 65536 00:27:15.009 } 00:27:15.009 ] 00:27:15.009 }' 00:27:15.009 13:47:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:15.009 13:47:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:15.267 [2024-11-20 13:47:22.192441] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:27:15.267 [2024-11-20 13:47:22.192537] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:27:15.267 [2024-11-20 13:47:22.242635] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:15.267 [2024-11-20 13:47:22.242680] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:15.267 [2024-11-20 13:47:22.242691] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 61291 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 61291 ']' 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 61291 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 61291 00:27:15.267 killing process with pid 61291 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:27:15.267 13:47:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 61291' 00:27:15.268 13:47:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 61291 00:27:15.268 [2024-11-20 13:47:22.303703] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:27:15.268 13:47:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 61291 00:27:15.268 [2024-11-20 13:47:22.312683] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:27:16.203 00:27:16.203 real 0m3.611s 00:27:16.203 user 0m5.195s 00:27:16.203 sys 0m0.632s 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:16.203 ************************************ 00:27:16.203 END TEST raid_state_function_test 00:27:16.203 ************************************ 00:27:16.203 13:47:22 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 2 true 00:27:16.203 13:47:22 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:27:16.203 13:47:22 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:27:16.203 13:47:22 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:27:16.203 ************************************ 00:27:16.203 START TEST raid_state_function_test_sb 00:27:16.203 ************************************ 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:27:16.203 Process raid pid: 61528 00:27:16.203 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=61528 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 61528' 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 61528 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 61528 ']' 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:16.203 13:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:27:16.203 [2024-11-20 13:47:23.036201] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:27:16.203 [2024-11-20 13:47:23.036485] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:27:16.203 [2024-11-20 13:47:23.190185] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:16.461 [2024-11-20 13:47:23.292381] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:27:16.461 [2024-11-20 13:47:23.415429] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:16.461 [2024-11-20 13:47:23.415469] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:17.025 13:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:17.025 13:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:27:17.025 13:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:27:17.025 13:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:17.025 13:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:17.025 [2024-11-20 13:47:23.897975] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:27:17.025 [2024-11-20 13:47:23.898038] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:27:17.025 [2024-11-20 13:47:23.898048] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:27:17.025 [2024-11-20 13:47:23.898056] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:27:17.025 13:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:17.025 13:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:27:17.025 13:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:17.025 13:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:17.025 13:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:17.025 13:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:17.025 13:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:17.025 13:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:17.025 13:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:17.025 13:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:17.025 13:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:17.025 13:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:17.025 13:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:17.025 13:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:17.025 13:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:17.025 13:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:17.025 13:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:17.025 "name": "Existed_Raid", 00:27:17.025 "uuid": "883562ff-65bf-44fc-aabd-9d0e3a153d82", 00:27:17.025 "strip_size_kb": 0, 00:27:17.025 "state": "configuring", 00:27:17.026 "raid_level": "raid1", 00:27:17.026 "superblock": true, 00:27:17.026 "num_base_bdevs": 2, 00:27:17.026 "num_base_bdevs_discovered": 0, 00:27:17.026 "num_base_bdevs_operational": 2, 00:27:17.026 "base_bdevs_list": [ 00:27:17.026 { 00:27:17.026 "name": "BaseBdev1", 00:27:17.026 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:17.026 "is_configured": false, 00:27:17.026 "data_offset": 0, 00:27:17.026 "data_size": 0 00:27:17.026 }, 00:27:17.026 { 00:27:17.026 "name": "BaseBdev2", 00:27:17.026 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:17.026 "is_configured": false, 00:27:17.026 "data_offset": 0, 00:27:17.026 "data_size": 0 00:27:17.026 } 00:27:17.026 ] 00:27:17.026 }' 00:27:17.026 13:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:17.026 13:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:17.284 [2024-11-20 13:47:24.222010] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:27:17.284 [2024-11-20 13:47:24.222042] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:17.284 [2024-11-20 13:47:24.229989] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:27:17.284 [2024-11-20 13:47:24.230029] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:27:17.284 [2024-11-20 13:47:24.230037] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:27:17.284 [2024-11-20 13:47:24.230047] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:17.284 [2024-11-20 13:47:24.260426] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:27:17.284 BaseBdev1 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:17.284 [ 00:27:17.284 { 00:27:17.284 "name": "BaseBdev1", 00:27:17.284 "aliases": [ 00:27:17.284 "a313c101-5019-480c-9ccb-ae764c7eae07" 00:27:17.284 ], 00:27:17.284 "product_name": "Malloc disk", 00:27:17.284 "block_size": 512, 00:27:17.284 "num_blocks": 65536, 00:27:17.284 "uuid": "a313c101-5019-480c-9ccb-ae764c7eae07", 00:27:17.284 "assigned_rate_limits": { 00:27:17.284 "rw_ios_per_sec": 0, 00:27:17.284 "rw_mbytes_per_sec": 0, 00:27:17.284 "r_mbytes_per_sec": 0, 00:27:17.284 "w_mbytes_per_sec": 0 00:27:17.284 }, 00:27:17.284 "claimed": true, 00:27:17.284 "claim_type": "exclusive_write", 00:27:17.284 "zoned": false, 00:27:17.284 "supported_io_types": { 00:27:17.284 "read": true, 00:27:17.284 "write": true, 00:27:17.284 "unmap": true, 00:27:17.284 "flush": true, 00:27:17.284 "reset": true, 00:27:17.284 "nvme_admin": false, 00:27:17.284 "nvme_io": false, 00:27:17.284 "nvme_io_md": false, 00:27:17.284 "write_zeroes": true, 00:27:17.284 "zcopy": true, 00:27:17.284 "get_zone_info": false, 00:27:17.284 "zone_management": false, 00:27:17.284 "zone_append": false, 00:27:17.284 "compare": false, 00:27:17.284 "compare_and_write": false, 00:27:17.284 "abort": true, 00:27:17.284 "seek_hole": false, 00:27:17.284 "seek_data": false, 00:27:17.284 "copy": true, 00:27:17.284 "nvme_iov_md": false 00:27:17.284 }, 00:27:17.284 "memory_domains": [ 00:27:17.284 { 00:27:17.284 "dma_device_id": "system", 00:27:17.284 "dma_device_type": 1 00:27:17.284 }, 00:27:17.284 { 00:27:17.284 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:17.284 "dma_device_type": 2 00:27:17.284 } 00:27:17.284 ], 00:27:17.284 "driver_specific": {} 00:27:17.284 } 00:27:17.284 ] 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:17.284 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:17.284 "name": "Existed_Raid", 00:27:17.284 "uuid": "188afb0a-fb62-4d51-bbbd-611228f31b51", 00:27:17.284 "strip_size_kb": 0, 00:27:17.284 "state": "configuring", 00:27:17.284 "raid_level": "raid1", 00:27:17.284 "superblock": true, 00:27:17.284 "num_base_bdevs": 2, 00:27:17.284 "num_base_bdevs_discovered": 1, 00:27:17.285 "num_base_bdevs_operational": 2, 00:27:17.285 "base_bdevs_list": [ 00:27:17.285 { 00:27:17.285 "name": "BaseBdev1", 00:27:17.285 "uuid": "a313c101-5019-480c-9ccb-ae764c7eae07", 00:27:17.285 "is_configured": true, 00:27:17.285 "data_offset": 2048, 00:27:17.285 "data_size": 63488 00:27:17.285 }, 00:27:17.285 { 00:27:17.285 "name": "BaseBdev2", 00:27:17.285 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:17.285 "is_configured": false, 00:27:17.285 "data_offset": 0, 00:27:17.285 "data_size": 0 00:27:17.285 } 00:27:17.285 ] 00:27:17.285 }' 00:27:17.285 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:17.285 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:17.852 [2024-11-20 13:47:24.612548] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:27:17.852 [2024-11-20 13:47:24.612599] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:17.852 [2024-11-20 13:47:24.620587] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:27:17.852 [2024-11-20 13:47:24.622296] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:27:17.852 [2024-11-20 13:47:24.622332] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:17.852 "name": "Existed_Raid", 00:27:17.852 "uuid": "a8fc76fb-9150-4347-aaf5-b80e5fa23271", 00:27:17.852 "strip_size_kb": 0, 00:27:17.852 "state": "configuring", 00:27:17.852 "raid_level": "raid1", 00:27:17.852 "superblock": true, 00:27:17.852 "num_base_bdevs": 2, 00:27:17.852 "num_base_bdevs_discovered": 1, 00:27:17.852 "num_base_bdevs_operational": 2, 00:27:17.852 "base_bdevs_list": [ 00:27:17.852 { 00:27:17.852 "name": "BaseBdev1", 00:27:17.852 "uuid": "a313c101-5019-480c-9ccb-ae764c7eae07", 00:27:17.852 "is_configured": true, 00:27:17.852 "data_offset": 2048, 00:27:17.852 "data_size": 63488 00:27:17.852 }, 00:27:17.852 { 00:27:17.852 "name": "BaseBdev2", 00:27:17.852 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:17.852 "is_configured": false, 00:27:17.852 "data_offset": 0, 00:27:17.852 "data_size": 0 00:27:17.852 } 00:27:17.852 ] 00:27:17.852 }' 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:17.852 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:18.126 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:27:18.126 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:18.126 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:18.126 [2024-11-20 13:47:24.969320] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:27:18.126 [2024-11-20 13:47:24.969525] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:27:18.126 [2024-11-20 13:47:24.969538] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:27:18.126 [2024-11-20 13:47:24.969762] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:27:18.126 BaseBdev2 00:27:18.126 [2024-11-20 13:47:24.969888] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:27:18.126 [2024-11-20 13:47:24.969899] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:27:18.126 [2024-11-20 13:47:24.970033] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:18.126 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:18.126 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:27:18.126 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:27:18.126 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:27:18.126 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:27:18.126 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:27:18.127 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:27:18.127 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:27:18.127 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:18.127 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:18.127 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:18.127 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:27:18.127 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:18.127 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:18.127 [ 00:27:18.127 { 00:27:18.127 "name": "BaseBdev2", 00:27:18.127 "aliases": [ 00:27:18.127 "84399586-5aa1-4d2f-a7ca-f0f171968bfe" 00:27:18.127 ], 00:27:18.127 "product_name": "Malloc disk", 00:27:18.127 "block_size": 512, 00:27:18.127 "num_blocks": 65536, 00:27:18.127 "uuid": "84399586-5aa1-4d2f-a7ca-f0f171968bfe", 00:27:18.127 "assigned_rate_limits": { 00:27:18.127 "rw_ios_per_sec": 0, 00:27:18.127 "rw_mbytes_per_sec": 0, 00:27:18.127 "r_mbytes_per_sec": 0, 00:27:18.127 "w_mbytes_per_sec": 0 00:27:18.127 }, 00:27:18.127 "claimed": true, 00:27:18.127 "claim_type": "exclusive_write", 00:27:18.127 "zoned": false, 00:27:18.127 "supported_io_types": { 00:27:18.127 "read": true, 00:27:18.127 "write": true, 00:27:18.127 "unmap": true, 00:27:18.127 "flush": true, 00:27:18.127 "reset": true, 00:27:18.127 "nvme_admin": false, 00:27:18.127 "nvme_io": false, 00:27:18.127 "nvme_io_md": false, 00:27:18.127 "write_zeroes": true, 00:27:18.127 "zcopy": true, 00:27:18.127 "get_zone_info": false, 00:27:18.127 "zone_management": false, 00:27:18.127 "zone_append": false, 00:27:18.127 "compare": false, 00:27:18.127 "compare_and_write": false, 00:27:18.127 "abort": true, 00:27:18.127 "seek_hole": false, 00:27:18.127 "seek_data": false, 00:27:18.127 "copy": true, 00:27:18.127 "nvme_iov_md": false 00:27:18.127 }, 00:27:18.127 "memory_domains": [ 00:27:18.127 { 00:27:18.127 "dma_device_id": "system", 00:27:18.127 "dma_device_type": 1 00:27:18.127 }, 00:27:18.127 { 00:27:18.127 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:18.127 "dma_device_type": 2 00:27:18.127 } 00:27:18.127 ], 00:27:18.127 "driver_specific": {} 00:27:18.127 } 00:27:18.127 ] 00:27:18.127 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:18.127 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:27:18.127 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:27:18.127 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:27:18.127 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:27:18.127 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:18.127 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:18.127 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:18.127 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:18.127 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:18.127 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:18.127 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:18.127 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:18.127 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:18.127 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:18.127 13:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:18.127 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:18.127 13:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:18.127 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:18.127 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:18.127 "name": "Existed_Raid", 00:27:18.127 "uuid": "a8fc76fb-9150-4347-aaf5-b80e5fa23271", 00:27:18.127 "strip_size_kb": 0, 00:27:18.127 "state": "online", 00:27:18.127 "raid_level": "raid1", 00:27:18.127 "superblock": true, 00:27:18.127 "num_base_bdevs": 2, 00:27:18.127 "num_base_bdevs_discovered": 2, 00:27:18.127 "num_base_bdevs_operational": 2, 00:27:18.127 "base_bdevs_list": [ 00:27:18.127 { 00:27:18.127 "name": "BaseBdev1", 00:27:18.127 "uuid": "a313c101-5019-480c-9ccb-ae764c7eae07", 00:27:18.127 "is_configured": true, 00:27:18.127 "data_offset": 2048, 00:27:18.127 "data_size": 63488 00:27:18.127 }, 00:27:18.127 { 00:27:18.127 "name": "BaseBdev2", 00:27:18.127 "uuid": "84399586-5aa1-4d2f-a7ca-f0f171968bfe", 00:27:18.127 "is_configured": true, 00:27:18.127 "data_offset": 2048, 00:27:18.127 "data_size": 63488 00:27:18.127 } 00:27:18.127 ] 00:27:18.127 }' 00:27:18.127 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:18.127 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:18.386 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:27:18.386 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:27:18.386 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:27:18.386 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:27:18.386 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:27:18.386 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:27:18.386 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:27:18.386 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:27:18.386 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:18.386 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:18.386 [2024-11-20 13:47:25.329710] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:18.386 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:18.386 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:27:18.386 "name": "Existed_Raid", 00:27:18.386 "aliases": [ 00:27:18.386 "a8fc76fb-9150-4347-aaf5-b80e5fa23271" 00:27:18.386 ], 00:27:18.386 "product_name": "Raid Volume", 00:27:18.386 "block_size": 512, 00:27:18.386 "num_blocks": 63488, 00:27:18.386 "uuid": "a8fc76fb-9150-4347-aaf5-b80e5fa23271", 00:27:18.386 "assigned_rate_limits": { 00:27:18.386 "rw_ios_per_sec": 0, 00:27:18.386 "rw_mbytes_per_sec": 0, 00:27:18.386 "r_mbytes_per_sec": 0, 00:27:18.386 "w_mbytes_per_sec": 0 00:27:18.386 }, 00:27:18.386 "claimed": false, 00:27:18.386 "zoned": false, 00:27:18.386 "supported_io_types": { 00:27:18.386 "read": true, 00:27:18.386 "write": true, 00:27:18.386 "unmap": false, 00:27:18.386 "flush": false, 00:27:18.386 "reset": true, 00:27:18.386 "nvme_admin": false, 00:27:18.386 "nvme_io": false, 00:27:18.386 "nvme_io_md": false, 00:27:18.386 "write_zeroes": true, 00:27:18.386 "zcopy": false, 00:27:18.386 "get_zone_info": false, 00:27:18.386 "zone_management": false, 00:27:18.386 "zone_append": false, 00:27:18.386 "compare": false, 00:27:18.386 "compare_and_write": false, 00:27:18.386 "abort": false, 00:27:18.386 "seek_hole": false, 00:27:18.386 "seek_data": false, 00:27:18.386 "copy": false, 00:27:18.386 "nvme_iov_md": false 00:27:18.386 }, 00:27:18.386 "memory_domains": [ 00:27:18.386 { 00:27:18.386 "dma_device_id": "system", 00:27:18.386 "dma_device_type": 1 00:27:18.386 }, 00:27:18.386 { 00:27:18.386 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:18.386 "dma_device_type": 2 00:27:18.386 }, 00:27:18.386 { 00:27:18.386 "dma_device_id": "system", 00:27:18.386 "dma_device_type": 1 00:27:18.386 }, 00:27:18.386 { 00:27:18.386 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:18.386 "dma_device_type": 2 00:27:18.386 } 00:27:18.386 ], 00:27:18.386 "driver_specific": { 00:27:18.386 "raid": { 00:27:18.386 "uuid": "a8fc76fb-9150-4347-aaf5-b80e5fa23271", 00:27:18.386 "strip_size_kb": 0, 00:27:18.386 "state": "online", 00:27:18.386 "raid_level": "raid1", 00:27:18.386 "superblock": true, 00:27:18.386 "num_base_bdevs": 2, 00:27:18.386 "num_base_bdevs_discovered": 2, 00:27:18.386 "num_base_bdevs_operational": 2, 00:27:18.386 "base_bdevs_list": [ 00:27:18.386 { 00:27:18.386 "name": "BaseBdev1", 00:27:18.386 "uuid": "a313c101-5019-480c-9ccb-ae764c7eae07", 00:27:18.386 "is_configured": true, 00:27:18.386 "data_offset": 2048, 00:27:18.386 "data_size": 63488 00:27:18.386 }, 00:27:18.386 { 00:27:18.386 "name": "BaseBdev2", 00:27:18.387 "uuid": "84399586-5aa1-4d2f-a7ca-f0f171968bfe", 00:27:18.387 "is_configured": true, 00:27:18.387 "data_offset": 2048, 00:27:18.387 "data_size": 63488 00:27:18.387 } 00:27:18.387 ] 00:27:18.387 } 00:27:18.387 } 00:27:18.387 }' 00:27:18.387 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:27:18.387 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:27:18.387 BaseBdev2' 00:27:18.387 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:18.387 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:27:18.387 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:18.387 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:27:18.387 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:18.387 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:18.387 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:18.645 [2024-11-20 13:47:25.501516] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:18.645 "name": "Existed_Raid", 00:27:18.645 "uuid": "a8fc76fb-9150-4347-aaf5-b80e5fa23271", 00:27:18.645 "strip_size_kb": 0, 00:27:18.645 "state": "online", 00:27:18.645 "raid_level": "raid1", 00:27:18.645 "superblock": true, 00:27:18.645 "num_base_bdevs": 2, 00:27:18.645 "num_base_bdevs_discovered": 1, 00:27:18.645 "num_base_bdevs_operational": 1, 00:27:18.645 "base_bdevs_list": [ 00:27:18.645 { 00:27:18.645 "name": null, 00:27:18.645 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:18.645 "is_configured": false, 00:27:18.645 "data_offset": 0, 00:27:18.645 "data_size": 63488 00:27:18.645 }, 00:27:18.645 { 00:27:18.645 "name": "BaseBdev2", 00:27:18.645 "uuid": "84399586-5aa1-4d2f-a7ca-f0f171968bfe", 00:27:18.645 "is_configured": true, 00:27:18.645 "data_offset": 2048, 00:27:18.645 "data_size": 63488 00:27:18.645 } 00:27:18.645 ] 00:27:18.645 }' 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:18.645 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:18.903 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:27:18.903 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:27:18.903 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:18.903 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:27:18.903 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:18.903 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:18.903 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:18.903 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:27:18.903 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:27:18.903 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:27:18.903 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:18.903 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:18.903 [2024-11-20 13:47:25.879335] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:27:18.903 [2024-11-20 13:47:25.879461] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:27:18.903 [2024-11-20 13:47:25.929970] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:18.903 [2024-11-20 13:47:25.930221] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:18.903 [2024-11-20 13:47:25.930241] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:27:18.903 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:18.903 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:27:18.903 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:27:18.903 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:27:18.903 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:18.903 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:18.903 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:18.903 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:19.160 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:27:19.160 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:27:19.160 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:27:19.160 13:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 61528 00:27:19.160 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 61528 ']' 00:27:19.160 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 61528 00:27:19.160 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:27:19.160 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:19.161 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 61528 00:27:19.161 killing process with pid 61528 00:27:19.161 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:27:19.161 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:27:19.161 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 61528' 00:27:19.161 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 61528 00:27:19.161 [2024-11-20 13:47:25.995842] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:27:19.161 13:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 61528 00:27:19.161 [2024-11-20 13:47:26.004849] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:27:19.730 13:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:27:19.730 00:27:19.730 real 0m3.655s 00:27:19.730 user 0m5.310s 00:27:19.730 sys 0m0.599s 00:27:19.730 13:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:27:19.730 ************************************ 00:27:19.730 END TEST raid_state_function_test_sb 00:27:19.730 ************************************ 00:27:19.730 13:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:19.730 13:47:26 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 2 00:27:19.730 13:47:26 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:27:19.730 13:47:26 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:27:19.730 13:47:26 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:27:19.730 ************************************ 00:27:19.730 START TEST raid_superblock_test 00:27:19.730 ************************************ 00:27:19.730 13:47:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:27:19.730 13:47:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:27:19.730 13:47:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:27:19.730 13:47:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:27:19.730 13:47:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:27:19.730 13:47:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:27:19.730 13:47:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:27:19.730 13:47:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:27:19.730 13:47:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:27:19.730 13:47:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:27:19.730 13:47:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:27:19.730 13:47:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:27:19.730 13:47:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:27:19.730 13:47:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:27:19.730 13:47:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:27:19.730 13:47:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:27:19.730 13:47:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=61769 00:27:19.730 13:47:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 61769 00:27:19.730 13:47:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 61769 ']' 00:27:19.730 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:19.730 13:47:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:19.730 13:47:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:19.730 13:47:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:27:19.730 13:47:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:19.730 13:47:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:19.730 13:47:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:19.730 [2024-11-20 13:47:26.743356] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:27:19.730 [2024-11-20 13:47:26.743497] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid61769 ] 00:27:19.987 [2024-11-20 13:47:26.898647] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:19.988 [2024-11-20 13:47:26.999411] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:27:20.245 [2024-11-20 13:47:27.120139] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:20.245 [2024-11-20 13:47:27.120179] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:20.503 13:47:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:20.503 13:47:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:27:20.503 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:27:20.503 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:27:20.503 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:27:20.503 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:27:20.503 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:27:20.503 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:27:20.503 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:27:20.503 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:27:20.503 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:27:20.503 13:47:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:20.503 13:47:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:20.761 malloc1 00:27:20.761 13:47:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:20.761 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:27:20.761 13:47:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:20.761 13:47:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:20.761 [2024-11-20 13:47:27.578817] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:27:20.761 [2024-11-20 13:47:27.578876] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:20.761 [2024-11-20 13:47:27.578895] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:27:20.761 [2024-11-20 13:47:27.578904] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:20.761 [2024-11-20 13:47:27.580877] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:20.761 [2024-11-20 13:47:27.580911] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:27:20.761 pt1 00:27:20.761 13:47:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:20.761 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:27:20.761 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:27:20.761 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:27:20.761 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:27:20.761 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:27:20.761 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:27:20.761 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:27:20.761 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:27:20.761 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:27:20.761 13:47:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:20.761 13:47:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:20.761 malloc2 00:27:20.761 13:47:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:20.761 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:27:20.761 13:47:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:20.761 13:47:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:20.761 [2024-11-20 13:47:27.612615] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:27:20.762 [2024-11-20 13:47:27.612661] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:20.762 [2024-11-20 13:47:27.612682] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:27:20.762 [2024-11-20 13:47:27.612690] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:20.762 [2024-11-20 13:47:27.614627] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:20.762 [2024-11-20 13:47:27.614784] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:27:20.762 pt2 00:27:20.762 13:47:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:20.762 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:27:20.762 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:27:20.762 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:27:20.762 13:47:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:20.762 13:47:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:20.762 [2024-11-20 13:47:27.620663] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:27:20.762 [2024-11-20 13:47:27.622364] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:27:20.762 [2024-11-20 13:47:27.622499] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:27:20.762 [2024-11-20 13:47:27.622513] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:27:20.762 [2024-11-20 13:47:27.622735] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:27:20.762 [2024-11-20 13:47:27.622857] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:27:20.762 [2024-11-20 13:47:27.622869] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:27:20.762 [2024-11-20 13:47:27.623007] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:20.762 13:47:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:20.762 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:27:20.762 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:20.762 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:20.762 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:20.762 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:20.762 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:20.762 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:20.762 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:20.762 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:20.762 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:20.762 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:20.762 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:20.762 13:47:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:20.762 13:47:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:20.762 13:47:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:20.762 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:20.762 "name": "raid_bdev1", 00:27:20.762 "uuid": "82f8b1fe-6a73-40b6-a9e1-c4d97dfbbdd2", 00:27:20.762 "strip_size_kb": 0, 00:27:20.762 "state": "online", 00:27:20.762 "raid_level": "raid1", 00:27:20.762 "superblock": true, 00:27:20.762 "num_base_bdevs": 2, 00:27:20.762 "num_base_bdevs_discovered": 2, 00:27:20.762 "num_base_bdevs_operational": 2, 00:27:20.762 "base_bdevs_list": [ 00:27:20.762 { 00:27:20.762 "name": "pt1", 00:27:20.762 "uuid": "00000000-0000-0000-0000-000000000001", 00:27:20.762 "is_configured": true, 00:27:20.762 "data_offset": 2048, 00:27:20.762 "data_size": 63488 00:27:20.762 }, 00:27:20.762 { 00:27:20.762 "name": "pt2", 00:27:20.762 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:20.762 "is_configured": true, 00:27:20.762 "data_offset": 2048, 00:27:20.762 "data_size": 63488 00:27:20.762 } 00:27:20.762 ] 00:27:20.762 }' 00:27:20.762 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:20.762 13:47:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:21.020 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:27:21.020 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:27:21.020 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:27:21.020 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:27:21.020 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:27:21.020 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:27:21.020 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:27:21.020 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:27:21.020 13:47:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:21.020 13:47:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:21.020 [2024-11-20 13:47:27.965002] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:21.020 13:47:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:21.020 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:27:21.020 "name": "raid_bdev1", 00:27:21.020 "aliases": [ 00:27:21.020 "82f8b1fe-6a73-40b6-a9e1-c4d97dfbbdd2" 00:27:21.020 ], 00:27:21.020 "product_name": "Raid Volume", 00:27:21.020 "block_size": 512, 00:27:21.020 "num_blocks": 63488, 00:27:21.020 "uuid": "82f8b1fe-6a73-40b6-a9e1-c4d97dfbbdd2", 00:27:21.020 "assigned_rate_limits": { 00:27:21.020 "rw_ios_per_sec": 0, 00:27:21.020 "rw_mbytes_per_sec": 0, 00:27:21.020 "r_mbytes_per_sec": 0, 00:27:21.020 "w_mbytes_per_sec": 0 00:27:21.020 }, 00:27:21.020 "claimed": false, 00:27:21.020 "zoned": false, 00:27:21.020 "supported_io_types": { 00:27:21.020 "read": true, 00:27:21.020 "write": true, 00:27:21.020 "unmap": false, 00:27:21.020 "flush": false, 00:27:21.020 "reset": true, 00:27:21.020 "nvme_admin": false, 00:27:21.020 "nvme_io": false, 00:27:21.020 "nvme_io_md": false, 00:27:21.020 "write_zeroes": true, 00:27:21.020 "zcopy": false, 00:27:21.020 "get_zone_info": false, 00:27:21.020 "zone_management": false, 00:27:21.020 "zone_append": false, 00:27:21.020 "compare": false, 00:27:21.020 "compare_and_write": false, 00:27:21.020 "abort": false, 00:27:21.020 "seek_hole": false, 00:27:21.020 "seek_data": false, 00:27:21.020 "copy": false, 00:27:21.020 "nvme_iov_md": false 00:27:21.020 }, 00:27:21.020 "memory_domains": [ 00:27:21.020 { 00:27:21.020 "dma_device_id": "system", 00:27:21.020 "dma_device_type": 1 00:27:21.020 }, 00:27:21.020 { 00:27:21.020 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:21.020 "dma_device_type": 2 00:27:21.020 }, 00:27:21.020 { 00:27:21.020 "dma_device_id": "system", 00:27:21.020 "dma_device_type": 1 00:27:21.020 }, 00:27:21.020 { 00:27:21.020 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:21.020 "dma_device_type": 2 00:27:21.020 } 00:27:21.020 ], 00:27:21.020 "driver_specific": { 00:27:21.020 "raid": { 00:27:21.020 "uuid": "82f8b1fe-6a73-40b6-a9e1-c4d97dfbbdd2", 00:27:21.020 "strip_size_kb": 0, 00:27:21.020 "state": "online", 00:27:21.020 "raid_level": "raid1", 00:27:21.020 "superblock": true, 00:27:21.020 "num_base_bdevs": 2, 00:27:21.020 "num_base_bdevs_discovered": 2, 00:27:21.020 "num_base_bdevs_operational": 2, 00:27:21.020 "base_bdevs_list": [ 00:27:21.020 { 00:27:21.020 "name": "pt1", 00:27:21.020 "uuid": "00000000-0000-0000-0000-000000000001", 00:27:21.020 "is_configured": true, 00:27:21.020 "data_offset": 2048, 00:27:21.020 "data_size": 63488 00:27:21.020 }, 00:27:21.020 { 00:27:21.020 "name": "pt2", 00:27:21.020 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:21.020 "is_configured": true, 00:27:21.020 "data_offset": 2048, 00:27:21.020 "data_size": 63488 00:27:21.020 } 00:27:21.020 ] 00:27:21.020 } 00:27:21.020 } 00:27:21.020 }' 00:27:21.020 13:47:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:27:21.020 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:27:21.020 pt2' 00:27:21.020 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:21.020 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:27:21.020 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:21.020 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:21.020 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:27:21.020 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:21.020 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:21.279 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:21.279 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:21.279 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:21.279 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:21.279 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:21.279 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:27:21.279 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:21.279 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:21.279 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:21.279 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:21.279 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:21.279 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:27:21.279 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:21.279 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:21.279 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:27:21.279 [2024-11-20 13:47:28.133010] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:21.279 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:21.279 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=82f8b1fe-6a73-40b6-a9e1-c4d97dfbbdd2 00:27:21.279 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 82f8b1fe-6a73-40b6-a9e1-c4d97dfbbdd2 ']' 00:27:21.279 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:27:21.279 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:21.279 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:21.279 [2024-11-20 13:47:28.164724] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:27:21.279 [2024-11-20 13:47:28.164747] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:27:21.279 [2024-11-20 13:47:28.164826] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:21.280 [2024-11-20 13:47:28.164882] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:21.280 [2024-11-20 13:47:28.164894] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:21.280 [2024-11-20 13:47:28.256764] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:27:21.280 [2024-11-20 13:47:28.258486] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:27:21.280 [2024-11-20 13:47:28.258544] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:27:21.280 [2024-11-20 13:47:28.258589] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:27:21.280 [2024-11-20 13:47:28.258602] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:27:21.280 [2024-11-20 13:47:28.258612] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:27:21.280 request: 00:27:21.280 { 00:27:21.280 "name": "raid_bdev1", 00:27:21.280 "raid_level": "raid1", 00:27:21.280 "base_bdevs": [ 00:27:21.280 "malloc1", 00:27:21.280 "malloc2" 00:27:21.280 ], 00:27:21.280 "superblock": false, 00:27:21.280 "method": "bdev_raid_create", 00:27:21.280 "req_id": 1 00:27:21.280 } 00:27:21.280 Got JSON-RPC error response 00:27:21.280 response: 00:27:21.280 { 00:27:21.280 "code": -17, 00:27:21.280 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:27:21.280 } 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:21.280 [2024-11-20 13:47:28.300762] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:27:21.280 [2024-11-20 13:47:28.300816] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:21.280 [2024-11-20 13:47:28.300834] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:27:21.280 [2024-11-20 13:47:28.300853] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:21.280 [2024-11-20 13:47:28.302862] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:21.280 [2024-11-20 13:47:28.303002] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:27:21.280 [2024-11-20 13:47:28.303085] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:27:21.280 [2024-11-20 13:47:28.303136] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:27:21.280 pt1 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:21.280 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:21.538 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:21.538 "name": "raid_bdev1", 00:27:21.538 "uuid": "82f8b1fe-6a73-40b6-a9e1-c4d97dfbbdd2", 00:27:21.538 "strip_size_kb": 0, 00:27:21.538 "state": "configuring", 00:27:21.538 "raid_level": "raid1", 00:27:21.538 "superblock": true, 00:27:21.538 "num_base_bdevs": 2, 00:27:21.538 "num_base_bdevs_discovered": 1, 00:27:21.538 "num_base_bdevs_operational": 2, 00:27:21.538 "base_bdevs_list": [ 00:27:21.538 { 00:27:21.538 "name": "pt1", 00:27:21.538 "uuid": "00000000-0000-0000-0000-000000000001", 00:27:21.538 "is_configured": true, 00:27:21.538 "data_offset": 2048, 00:27:21.538 "data_size": 63488 00:27:21.538 }, 00:27:21.538 { 00:27:21.538 "name": null, 00:27:21.538 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:21.538 "is_configured": false, 00:27:21.538 "data_offset": 2048, 00:27:21.538 "data_size": 63488 00:27:21.538 } 00:27:21.538 ] 00:27:21.538 }' 00:27:21.539 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:21.539 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:21.797 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:27:21.797 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:27:21.797 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:27:21.797 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:27:21.797 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:21.797 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:21.797 [2024-11-20 13:47:28.640858] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:27:21.797 [2024-11-20 13:47:28.640924] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:21.797 [2024-11-20 13:47:28.640943] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:27:21.797 [2024-11-20 13:47:28.640953] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:21.797 [2024-11-20 13:47:28.641378] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:21.797 [2024-11-20 13:47:28.641397] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:27:21.797 [2024-11-20 13:47:28.641470] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:27:21.797 [2024-11-20 13:47:28.641494] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:27:21.797 [2024-11-20 13:47:28.641595] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:27:21.797 [2024-11-20 13:47:28.641605] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:27:21.797 [2024-11-20 13:47:28.641814] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:27:21.797 [2024-11-20 13:47:28.641934] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:27:21.797 [2024-11-20 13:47:28.641941] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:27:21.797 [2024-11-20 13:47:28.642097] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:21.797 pt2 00:27:21.797 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:21.797 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:27:21.797 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:27:21.797 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:27:21.797 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:21.797 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:21.797 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:21.797 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:21.797 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:21.797 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:21.797 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:21.797 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:21.797 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:21.797 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:21.797 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:21.797 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:21.797 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:21.797 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:21.797 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:21.797 "name": "raid_bdev1", 00:27:21.797 "uuid": "82f8b1fe-6a73-40b6-a9e1-c4d97dfbbdd2", 00:27:21.797 "strip_size_kb": 0, 00:27:21.797 "state": "online", 00:27:21.797 "raid_level": "raid1", 00:27:21.797 "superblock": true, 00:27:21.798 "num_base_bdevs": 2, 00:27:21.798 "num_base_bdevs_discovered": 2, 00:27:21.798 "num_base_bdevs_operational": 2, 00:27:21.798 "base_bdevs_list": [ 00:27:21.798 { 00:27:21.798 "name": "pt1", 00:27:21.798 "uuid": "00000000-0000-0000-0000-000000000001", 00:27:21.798 "is_configured": true, 00:27:21.798 "data_offset": 2048, 00:27:21.798 "data_size": 63488 00:27:21.798 }, 00:27:21.798 { 00:27:21.798 "name": "pt2", 00:27:21.798 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:21.798 "is_configured": true, 00:27:21.798 "data_offset": 2048, 00:27:21.798 "data_size": 63488 00:27:21.798 } 00:27:21.798 ] 00:27:21.798 }' 00:27:21.798 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:21.798 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:22.056 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:27:22.056 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:27:22.056 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:27:22.056 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:27:22.056 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:27:22.056 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:27:22.056 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:27:22.056 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:27:22.056 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:22.056 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:22.056 [2024-11-20 13:47:28.941164] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:22.056 13:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:22.056 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:27:22.056 "name": "raid_bdev1", 00:27:22.056 "aliases": [ 00:27:22.056 "82f8b1fe-6a73-40b6-a9e1-c4d97dfbbdd2" 00:27:22.056 ], 00:27:22.056 "product_name": "Raid Volume", 00:27:22.056 "block_size": 512, 00:27:22.056 "num_blocks": 63488, 00:27:22.056 "uuid": "82f8b1fe-6a73-40b6-a9e1-c4d97dfbbdd2", 00:27:22.056 "assigned_rate_limits": { 00:27:22.056 "rw_ios_per_sec": 0, 00:27:22.056 "rw_mbytes_per_sec": 0, 00:27:22.056 "r_mbytes_per_sec": 0, 00:27:22.056 "w_mbytes_per_sec": 0 00:27:22.056 }, 00:27:22.056 "claimed": false, 00:27:22.056 "zoned": false, 00:27:22.056 "supported_io_types": { 00:27:22.056 "read": true, 00:27:22.056 "write": true, 00:27:22.056 "unmap": false, 00:27:22.056 "flush": false, 00:27:22.056 "reset": true, 00:27:22.056 "nvme_admin": false, 00:27:22.056 "nvme_io": false, 00:27:22.056 "nvme_io_md": false, 00:27:22.056 "write_zeroes": true, 00:27:22.056 "zcopy": false, 00:27:22.056 "get_zone_info": false, 00:27:22.056 "zone_management": false, 00:27:22.056 "zone_append": false, 00:27:22.056 "compare": false, 00:27:22.056 "compare_and_write": false, 00:27:22.056 "abort": false, 00:27:22.056 "seek_hole": false, 00:27:22.056 "seek_data": false, 00:27:22.057 "copy": false, 00:27:22.057 "nvme_iov_md": false 00:27:22.057 }, 00:27:22.057 "memory_domains": [ 00:27:22.057 { 00:27:22.057 "dma_device_id": "system", 00:27:22.057 "dma_device_type": 1 00:27:22.057 }, 00:27:22.057 { 00:27:22.057 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:22.057 "dma_device_type": 2 00:27:22.057 }, 00:27:22.057 { 00:27:22.057 "dma_device_id": "system", 00:27:22.057 "dma_device_type": 1 00:27:22.057 }, 00:27:22.057 { 00:27:22.057 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:22.057 "dma_device_type": 2 00:27:22.057 } 00:27:22.057 ], 00:27:22.057 "driver_specific": { 00:27:22.057 "raid": { 00:27:22.057 "uuid": "82f8b1fe-6a73-40b6-a9e1-c4d97dfbbdd2", 00:27:22.057 "strip_size_kb": 0, 00:27:22.057 "state": "online", 00:27:22.057 "raid_level": "raid1", 00:27:22.057 "superblock": true, 00:27:22.057 "num_base_bdevs": 2, 00:27:22.057 "num_base_bdevs_discovered": 2, 00:27:22.057 "num_base_bdevs_operational": 2, 00:27:22.057 "base_bdevs_list": [ 00:27:22.057 { 00:27:22.057 "name": "pt1", 00:27:22.057 "uuid": "00000000-0000-0000-0000-000000000001", 00:27:22.057 "is_configured": true, 00:27:22.057 "data_offset": 2048, 00:27:22.057 "data_size": 63488 00:27:22.057 }, 00:27:22.057 { 00:27:22.057 "name": "pt2", 00:27:22.057 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:22.057 "is_configured": true, 00:27:22.057 "data_offset": 2048, 00:27:22.057 "data_size": 63488 00:27:22.057 } 00:27:22.057 ] 00:27:22.057 } 00:27:22.057 } 00:27:22.057 }' 00:27:22.057 13:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:27:22.057 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:27:22.057 pt2' 00:27:22.057 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:22.057 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:27:22.057 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:22.057 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:27:22.057 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:22.057 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:22.057 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:22.057 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:22.057 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:22.057 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:22.057 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:22.057 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:22.057 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:27:22.057 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:22.057 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:22.057 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:22.057 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:22.057 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:22.057 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:27:22.057 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:22.057 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:22.057 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:27:22.315 [2024-11-20 13:47:29.113175] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 82f8b1fe-6a73-40b6-a9e1-c4d97dfbbdd2 '!=' 82f8b1fe-6a73-40b6-a9e1-c4d97dfbbdd2 ']' 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:22.315 [2024-11-20 13:47:29.145005] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:22.315 "name": "raid_bdev1", 00:27:22.315 "uuid": "82f8b1fe-6a73-40b6-a9e1-c4d97dfbbdd2", 00:27:22.315 "strip_size_kb": 0, 00:27:22.315 "state": "online", 00:27:22.315 "raid_level": "raid1", 00:27:22.315 "superblock": true, 00:27:22.315 "num_base_bdevs": 2, 00:27:22.315 "num_base_bdevs_discovered": 1, 00:27:22.315 "num_base_bdevs_operational": 1, 00:27:22.315 "base_bdevs_list": [ 00:27:22.315 { 00:27:22.315 "name": null, 00:27:22.315 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:22.315 "is_configured": false, 00:27:22.315 "data_offset": 0, 00:27:22.315 "data_size": 63488 00:27:22.315 }, 00:27:22.315 { 00:27:22.315 "name": "pt2", 00:27:22.315 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:22.315 "is_configured": true, 00:27:22.315 "data_offset": 2048, 00:27:22.315 "data_size": 63488 00:27:22.315 } 00:27:22.315 ] 00:27:22.315 }' 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:22.315 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:22.574 [2024-11-20 13:47:29.477027] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:27:22.574 [2024-11-20 13:47:29.477055] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:27:22.574 [2024-11-20 13:47:29.477130] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:22.574 [2024-11-20 13:47:29.477177] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:22.574 [2024-11-20 13:47:29.477187] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=1 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:22.574 [2024-11-20 13:47:29.525012] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:27:22.574 [2024-11-20 13:47:29.525142] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:22.574 [2024-11-20 13:47:29.525175] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:27:22.574 [2024-11-20 13:47:29.525221] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:22.574 [2024-11-20 13:47:29.527280] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:22.574 [2024-11-20 13:47:29.527379] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:27:22.574 [2024-11-20 13:47:29.527509] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:27:22.574 [2024-11-20 13:47:29.527570] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:27:22.574 [2024-11-20 13:47:29.527732] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:27:22.574 [2024-11-20 13:47:29.527757] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:27:22.574 [2024-11-20 13:47:29.527994] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:27:22.574 [2024-11-20 13:47:29.528181] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:27:22.574 [2024-11-20 13:47:29.528239] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:27:22.574 [2024-11-20 13:47:29.528410] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:22.574 pt2 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:22.574 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:22.575 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:22.575 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:27:22.575 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:22.575 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:22.575 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:22.575 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:22.575 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:22.575 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:22.575 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:22.575 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:22.575 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:22.575 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:22.575 "name": "raid_bdev1", 00:27:22.575 "uuid": "82f8b1fe-6a73-40b6-a9e1-c4d97dfbbdd2", 00:27:22.575 "strip_size_kb": 0, 00:27:22.575 "state": "online", 00:27:22.575 "raid_level": "raid1", 00:27:22.575 "superblock": true, 00:27:22.575 "num_base_bdevs": 2, 00:27:22.575 "num_base_bdevs_discovered": 1, 00:27:22.575 "num_base_bdevs_operational": 1, 00:27:22.575 "base_bdevs_list": [ 00:27:22.575 { 00:27:22.575 "name": null, 00:27:22.575 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:22.575 "is_configured": false, 00:27:22.575 "data_offset": 2048, 00:27:22.575 "data_size": 63488 00:27:22.575 }, 00:27:22.575 { 00:27:22.575 "name": "pt2", 00:27:22.575 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:22.575 "is_configured": true, 00:27:22.575 "data_offset": 2048, 00:27:22.575 "data_size": 63488 00:27:22.575 } 00:27:22.575 ] 00:27:22.575 }' 00:27:22.575 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:22.575 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:22.834 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:27:22.834 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:22.834 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:22.834 [2024-11-20 13:47:29.825055] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:27:22.834 [2024-11-20 13:47:29.825085] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:27:22.834 [2024-11-20 13:47:29.825156] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:22.834 [2024-11-20 13:47:29.825206] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:22.834 [2024-11-20 13:47:29.825214] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:27:22.834 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:22.834 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:27:22.834 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:22.834 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:22.834 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:22.834 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:22.834 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:27:22.834 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:27:22.834 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:27:22.834 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:27:22.834 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:22.834 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:22.834 [2024-11-20 13:47:29.865078] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:27:22.834 [2024-11-20 13:47:29.865218] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:22.834 [2024-11-20 13:47:29.865253] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:27:22.834 [2024-11-20 13:47:29.865437] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:22.834 [2024-11-20 13:47:29.867509] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:22.834 [2024-11-20 13:47:29.867597] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:27:22.834 [2024-11-20 13:47:29.867717] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:27:22.834 [2024-11-20 13:47:29.867798] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:27:22.834 [2024-11-20 13:47:29.867939] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:27:22.834 [2024-11-20 13:47:29.868081] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:27:22.834 [2024-11-20 13:47:29.868140] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:27:22.834 [2024-11-20 13:47:29.868236] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:27:22.834 [2024-11-20 13:47:29.868326] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:27:22.834 [2024-11-20 13:47:29.868373] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:27:22.834 [2024-11-20 13:47:29.868614] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:27:22.834 [2024-11-20 13:47:29.868781] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:27:22.834 [2024-11-20 13:47:29.868851] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:27:22.834 [2024-11-20 13:47:29.869091] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:22.834 pt1 00:27:22.834 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:22.835 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:27:22.835 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:27:22.835 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:22.835 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:22.835 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:22.835 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:22.835 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:27:22.835 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:22.835 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:22.835 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:22.835 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:22.835 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:22.835 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:22.835 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:22.835 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:22.835 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:23.093 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:23.093 "name": "raid_bdev1", 00:27:23.093 "uuid": "82f8b1fe-6a73-40b6-a9e1-c4d97dfbbdd2", 00:27:23.093 "strip_size_kb": 0, 00:27:23.093 "state": "online", 00:27:23.093 "raid_level": "raid1", 00:27:23.093 "superblock": true, 00:27:23.093 "num_base_bdevs": 2, 00:27:23.093 "num_base_bdevs_discovered": 1, 00:27:23.093 "num_base_bdevs_operational": 1, 00:27:23.093 "base_bdevs_list": [ 00:27:23.093 { 00:27:23.093 "name": null, 00:27:23.093 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:23.093 "is_configured": false, 00:27:23.093 "data_offset": 2048, 00:27:23.093 "data_size": 63488 00:27:23.093 }, 00:27:23.093 { 00:27:23.093 "name": "pt2", 00:27:23.093 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:23.093 "is_configured": true, 00:27:23.093 "data_offset": 2048, 00:27:23.093 "data_size": 63488 00:27:23.093 } 00:27:23.093 ] 00:27:23.093 }' 00:27:23.093 13:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:23.093 13:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:23.351 13:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:27:23.351 13:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:27:23.351 13:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:23.351 13:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:23.351 13:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:23.351 13:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:27:23.351 13:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:27:23.351 13:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:27:23.351 13:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:23.351 13:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:23.351 [2024-11-20 13:47:30.237415] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:23.351 13:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:23.351 13:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 82f8b1fe-6a73-40b6-a9e1-c4d97dfbbdd2 '!=' 82f8b1fe-6a73-40b6-a9e1-c4d97dfbbdd2 ']' 00:27:23.351 13:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 61769 00:27:23.351 13:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 61769 ']' 00:27:23.351 13:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 61769 00:27:23.351 13:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:27:23.351 13:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:23.351 13:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 61769 00:27:23.351 killing process with pid 61769 00:27:23.351 13:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:27:23.351 13:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:27:23.351 13:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 61769' 00:27:23.351 13:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 61769 00:27:23.351 13:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 61769 00:27:23.351 [2024-11-20 13:47:30.282295] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:27:23.351 [2024-11-20 13:47:30.282389] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:23.351 [2024-11-20 13:47:30.282433] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:23.351 [2024-11-20 13:47:30.282449] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:27:23.351 [2024-11-20 13:47:30.390029] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:27:24.285 13:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:27:24.285 00:27:24.285 real 0m4.321s 00:27:24.285 user 0m6.571s 00:27:24.285 sys 0m0.748s 00:27:24.285 13:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:27:24.285 ************************************ 00:27:24.285 END TEST raid_superblock_test 00:27:24.285 ************************************ 00:27:24.285 13:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:24.285 13:47:31 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 2 read 00:27:24.285 13:47:31 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:27:24.285 13:47:31 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:27:24.285 13:47:31 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:27:24.285 ************************************ 00:27:24.285 START TEST raid_read_error_test 00:27:24.285 ************************************ 00:27:24.285 13:47:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 2 read 00:27:24.285 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:27:24.285 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:27:24.285 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:27:24.285 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:27:24.285 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:27:24.285 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:27:24.285 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:27:24.285 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:27:24.285 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:27:24.285 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:27:24.285 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:27:24.285 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:27:24.285 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:27:24.285 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:27:24.285 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:27:24.285 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:27:24.285 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:27:24.285 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:27:24.285 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:27:24.286 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:27:24.286 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:27:24.286 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:24.286 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.N6Rc2UvBeS 00:27:24.286 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=62077 00:27:24.286 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 62077 00:27:24.286 13:47:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 62077 ']' 00:27:24.286 13:47:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:24.286 13:47:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:24.286 13:47:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:24.286 13:47:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:24.286 13:47:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:24.286 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:27:24.286 [2024-11-20 13:47:31.117083] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:27:24.286 [2024-11-20 13:47:31.117203] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid62077 ] 00:27:24.286 [2024-11-20 13:47:31.278759] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:24.544 [2024-11-20 13:47:31.382975] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:27:24.544 [2024-11-20 13:47:31.506050] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:24.544 [2024-11-20 13:47:31.506273] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:25.110 13:47:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:25.110 13:47:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:27:25.110 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:27:25.110 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:27:25.110 13:47:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:25.110 13:47:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:25.110 BaseBdev1_malloc 00:27:25.110 13:47:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:25.110 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:27:25.110 13:47:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:25.110 13:47:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:25.110 true 00:27:25.110 13:47:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:25.110 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:27:25.110 13:47:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:25.110 13:47:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:25.110 [2024-11-20 13:47:31.984195] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:27:25.110 [2024-11-20 13:47:31.984248] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:25.110 [2024-11-20 13:47:31.984266] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:27:25.110 [2024-11-20 13:47:31.984276] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:25.110 [2024-11-20 13:47:31.986210] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:25.110 [2024-11-20 13:47:31.986244] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:27:25.110 BaseBdev1 00:27:25.110 13:47:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:25.110 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:27:25.110 13:47:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:27:25.110 13:47:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:25.110 13:47:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:25.110 BaseBdev2_malloc 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:25.110 true 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:25.110 [2024-11-20 13:47:32.026227] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:27:25.110 [2024-11-20 13:47:32.026273] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:25.110 [2024-11-20 13:47:32.026286] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:27:25.110 [2024-11-20 13:47:32.026296] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:25.110 [2024-11-20 13:47:32.028242] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:25.110 [2024-11-20 13:47:32.028404] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:27:25.110 BaseBdev2 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:25.110 [2024-11-20 13:47:32.034282] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:27:25.110 [2024-11-20 13:47:32.035953] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:27:25.110 [2024-11-20 13:47:32.036231] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:27:25.110 [2024-11-20 13:47:32.036250] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:27:25.110 [2024-11-20 13:47:32.036467] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:27:25.110 [2024-11-20 13:47:32.036616] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:27:25.110 [2024-11-20 13:47:32.036626] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:27:25.110 [2024-11-20 13:47:32.036747] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:25.110 "name": "raid_bdev1", 00:27:25.110 "uuid": "284e83b7-73d9-4623-b289-1e9c2b6fe673", 00:27:25.110 "strip_size_kb": 0, 00:27:25.110 "state": "online", 00:27:25.110 "raid_level": "raid1", 00:27:25.110 "superblock": true, 00:27:25.110 "num_base_bdevs": 2, 00:27:25.110 "num_base_bdevs_discovered": 2, 00:27:25.110 "num_base_bdevs_operational": 2, 00:27:25.110 "base_bdevs_list": [ 00:27:25.110 { 00:27:25.110 "name": "BaseBdev1", 00:27:25.110 "uuid": "a8443699-be9f-56bc-86be-47a7f58202a8", 00:27:25.110 "is_configured": true, 00:27:25.110 "data_offset": 2048, 00:27:25.110 "data_size": 63488 00:27:25.110 }, 00:27:25.110 { 00:27:25.110 "name": "BaseBdev2", 00:27:25.110 "uuid": "dbf750ac-7346-5a80-b420-9f09a869523b", 00:27:25.110 "is_configured": true, 00:27:25.110 "data_offset": 2048, 00:27:25.110 "data_size": 63488 00:27:25.110 } 00:27:25.110 ] 00:27:25.110 }' 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:25.110 13:47:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:25.368 13:47:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:27:25.368 13:47:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:27:25.625 [2024-11-20 13:47:32.463285] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:27:26.561 13:47:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:27:26.561 13:47:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:26.561 13:47:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:26.561 13:47:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:26.561 13:47:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:27:26.561 13:47:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:27:26.561 13:47:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:27:26.561 13:47:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:27:26.561 13:47:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:27:26.561 13:47:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:26.561 13:47:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:26.561 13:47:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:26.561 13:47:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:26.561 13:47:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:26.561 13:47:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:26.561 13:47:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:26.561 13:47:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:26.561 13:47:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:26.561 13:47:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:26.561 13:47:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:26.561 13:47:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:26.561 13:47:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:26.561 13:47:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:26.561 13:47:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:26.561 "name": "raid_bdev1", 00:27:26.561 "uuid": "284e83b7-73d9-4623-b289-1e9c2b6fe673", 00:27:26.561 "strip_size_kb": 0, 00:27:26.561 "state": "online", 00:27:26.561 "raid_level": "raid1", 00:27:26.561 "superblock": true, 00:27:26.561 "num_base_bdevs": 2, 00:27:26.561 "num_base_bdevs_discovered": 2, 00:27:26.561 "num_base_bdevs_operational": 2, 00:27:26.561 "base_bdevs_list": [ 00:27:26.561 { 00:27:26.561 "name": "BaseBdev1", 00:27:26.561 "uuid": "a8443699-be9f-56bc-86be-47a7f58202a8", 00:27:26.561 "is_configured": true, 00:27:26.561 "data_offset": 2048, 00:27:26.561 "data_size": 63488 00:27:26.561 }, 00:27:26.561 { 00:27:26.561 "name": "BaseBdev2", 00:27:26.561 "uuid": "dbf750ac-7346-5a80-b420-9f09a869523b", 00:27:26.561 "is_configured": true, 00:27:26.561 "data_offset": 2048, 00:27:26.561 "data_size": 63488 00:27:26.561 } 00:27:26.561 ] 00:27:26.561 }' 00:27:26.561 13:47:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:26.561 13:47:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:26.819 13:47:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:27:26.819 13:47:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:26.819 13:47:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:26.819 [2024-11-20 13:47:33.682552] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:27:26.819 [2024-11-20 13:47:33.682746] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:27:26.819 [2024-11-20 13:47:33.685334] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:26.819 [2024-11-20 13:47:33.685378] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:26.819 [2024-11-20 13:47:33.685460] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:26.819 [2024-11-20 13:47:33.685471] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:27:26.819 13:47:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:26.819 { 00:27:26.819 "results": [ 00:27:26.819 { 00:27:26.819 "job": "raid_bdev1", 00:27:26.819 "core_mask": "0x1", 00:27:26.819 "workload": "randrw", 00:27:26.819 "percentage": 50, 00:27:26.819 "status": "finished", 00:27:26.819 "queue_depth": 1, 00:27:26.819 "io_size": 131072, 00:27:26.819 "runtime": 1.217577, 00:27:26.819 "iops": 19197.143178624432, 00:27:26.819 "mibps": 2399.642897328054, 00:27:26.819 "io_failed": 0, 00:27:26.819 "io_timeout": 0, 00:27:26.819 "avg_latency_us": 49.46211003679302, 00:27:26.819 "min_latency_us": 23.138461538461538, 00:27:26.819 "max_latency_us": 1392.64 00:27:26.819 } 00:27:26.819 ], 00:27:26.819 "core_count": 1 00:27:26.819 } 00:27:26.819 13:47:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 62077 00:27:26.819 13:47:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 62077 ']' 00:27:26.819 13:47:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 62077 00:27:26.819 13:47:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:27:26.819 13:47:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:26.819 13:47:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 62077 00:27:26.819 killing process with pid 62077 00:27:26.819 13:47:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:27:26.819 13:47:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:27:26.819 13:47:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 62077' 00:27:26.819 13:47:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 62077 00:27:26.819 [2024-11-20 13:47:33.715480] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:27:26.819 13:47:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 62077 00:27:26.819 [2024-11-20 13:47:33.788494] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:27:27.755 13:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:27:27.755 13:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.N6Rc2UvBeS 00:27:27.755 13:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:27:27.755 13:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:27:27.755 13:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:27:27.755 13:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:27:27.755 13:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:27:27.755 13:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:27:27.755 00:27:27.755 real 0m3.420s 00:27:27.755 user 0m4.096s 00:27:27.755 sys 0m0.413s 00:27:27.755 ************************************ 00:27:27.755 END TEST raid_read_error_test 00:27:27.755 ************************************ 00:27:27.755 13:47:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:27:27.755 13:47:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:27.755 13:47:34 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 2 write 00:27:27.755 13:47:34 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:27:27.755 13:47:34 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:27:27.755 13:47:34 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:27:27.755 ************************************ 00:27:27.755 START TEST raid_write_error_test 00:27:27.755 ************************************ 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 2 write 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.Khi4dcrfi8 00:27:27.755 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=62206 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 62206 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 62206 ']' 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:27.755 13:47:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:27:27.755 [2024-11-20 13:47:34.587944] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:27:27.755 [2024-11-20 13:47:34.588660] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid62206 ] 00:27:27.755 [2024-11-20 13:47:34.752409] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:28.014 [2024-11-20 13:47:34.871642] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:27:28.014 [2024-11-20 13:47:35.019478] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:28.014 [2024-11-20 13:47:35.019553] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:28.581 BaseBdev1_malloc 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:28.581 true 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:28.581 [2024-11-20 13:47:35.487526] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:27:28.581 [2024-11-20 13:47:35.487589] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:28.581 [2024-11-20 13:47:35.487610] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:27:28.581 [2024-11-20 13:47:35.487622] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:28.581 [2024-11-20 13:47:35.489850] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:28.581 [2024-11-20 13:47:35.489888] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:27:28.581 BaseBdev1 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:28.581 BaseBdev2_malloc 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:28.581 true 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:28.581 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:28.581 [2024-11-20 13:47:35.533704] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:27:28.581 [2024-11-20 13:47:35.533933] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:28.582 [2024-11-20 13:47:35.533958] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:27:28.582 [2024-11-20 13:47:35.533969] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:28.582 [2024-11-20 13:47:35.536235] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:28.582 [2024-11-20 13:47:35.536271] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:27:28.582 BaseBdev2 00:27:28.582 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:28.582 13:47:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:27:28.582 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:28.582 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:28.582 [2024-11-20 13:47:35.541771] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:27:28.582 [2024-11-20 13:47:35.543734] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:27:28.582 [2024-11-20 13:47:35.543917] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:27:28.582 [2024-11-20 13:47:35.543932] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:27:28.582 [2024-11-20 13:47:35.544197] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:27:28.582 [2024-11-20 13:47:35.544494] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:27:28.582 [2024-11-20 13:47:35.544509] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:27:28.582 [2024-11-20 13:47:35.544653] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:28.582 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:28.582 13:47:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:27:28.582 13:47:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:28.582 13:47:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:28.582 13:47:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:28.582 13:47:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:28.582 13:47:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:28.582 13:47:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:28.582 13:47:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:28.582 13:47:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:28.582 13:47:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:28.582 13:47:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:28.582 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:28.582 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:28.582 13:47:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:28.582 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:28.582 13:47:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:28.582 "name": "raid_bdev1", 00:27:28.582 "uuid": "79d6492f-8b8a-4b00-9b85-91287dc80694", 00:27:28.582 "strip_size_kb": 0, 00:27:28.582 "state": "online", 00:27:28.582 "raid_level": "raid1", 00:27:28.582 "superblock": true, 00:27:28.582 "num_base_bdevs": 2, 00:27:28.582 "num_base_bdevs_discovered": 2, 00:27:28.582 "num_base_bdevs_operational": 2, 00:27:28.582 "base_bdevs_list": [ 00:27:28.582 { 00:27:28.582 "name": "BaseBdev1", 00:27:28.582 "uuid": "06c0731a-f435-53ff-aed3-6e8a0f3360cd", 00:27:28.582 "is_configured": true, 00:27:28.582 "data_offset": 2048, 00:27:28.582 "data_size": 63488 00:27:28.582 }, 00:27:28.582 { 00:27:28.582 "name": "BaseBdev2", 00:27:28.582 "uuid": "4b7ab16a-99f2-56c7-97e0-0c480988fb8c", 00:27:28.582 "is_configured": true, 00:27:28.582 "data_offset": 2048, 00:27:28.582 "data_size": 63488 00:27:28.582 } 00:27:28.582 ] 00:27:28.582 }' 00:27:28.582 13:47:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:28.582 13:47:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:28.839 13:47:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:27:28.839 13:47:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:27:29.096 [2024-11-20 13:47:35.950871] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:27:30.030 13:47:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:27:30.030 13:47:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:30.030 13:47:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:30.030 [2024-11-20 13:47:36.869134] bdev_raid.c:2276:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:27:30.030 [2024-11-20 13:47:36.869203] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:27:30.030 [2024-11-20 13:47:36.869402] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d0000063c0 00:27:30.030 13:47:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:30.030 13:47:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:27:30.030 13:47:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:27:30.030 13:47:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:27:30.030 13:47:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=1 00:27:30.030 13:47:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:27:30.030 13:47:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:30.030 13:47:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:30.030 13:47:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:30.030 13:47:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:30.030 13:47:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:27:30.030 13:47:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:30.030 13:47:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:30.030 13:47:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:30.030 13:47:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:30.030 13:47:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:30.030 13:47:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:30.030 13:47:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:30.030 13:47:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:30.030 13:47:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:30.030 13:47:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:30.030 "name": "raid_bdev1", 00:27:30.030 "uuid": "79d6492f-8b8a-4b00-9b85-91287dc80694", 00:27:30.030 "strip_size_kb": 0, 00:27:30.030 "state": "online", 00:27:30.030 "raid_level": "raid1", 00:27:30.031 "superblock": true, 00:27:30.031 "num_base_bdevs": 2, 00:27:30.031 "num_base_bdevs_discovered": 1, 00:27:30.031 "num_base_bdevs_operational": 1, 00:27:30.031 "base_bdevs_list": [ 00:27:30.031 { 00:27:30.031 "name": null, 00:27:30.031 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:30.031 "is_configured": false, 00:27:30.031 "data_offset": 0, 00:27:30.031 "data_size": 63488 00:27:30.031 }, 00:27:30.031 { 00:27:30.031 "name": "BaseBdev2", 00:27:30.031 "uuid": "4b7ab16a-99f2-56c7-97e0-0c480988fb8c", 00:27:30.031 "is_configured": true, 00:27:30.031 "data_offset": 2048, 00:27:30.031 "data_size": 63488 00:27:30.031 } 00:27:30.031 ] 00:27:30.031 }' 00:27:30.031 13:47:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:30.031 13:47:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:30.289 13:47:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:27:30.289 13:47:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:30.289 13:47:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:30.289 [2024-11-20 13:47:37.183923] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:27:30.289 [2024-11-20 13:47:37.183958] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:27:30.289 [2024-11-20 13:47:37.187030] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:30.289 [2024-11-20 13:47:37.187073] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:30.289 [2024-11-20 13:47:37.187142] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:30.289 [2024-11-20 13:47:37.187154] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:27:30.289 { 00:27:30.289 "results": [ 00:27:30.289 { 00:27:30.289 "job": "raid_bdev1", 00:27:30.289 "core_mask": "0x1", 00:27:30.289 "workload": "randrw", 00:27:30.289 "percentage": 50, 00:27:30.289 "status": "finished", 00:27:30.289 "queue_depth": 1, 00:27:30.289 "io_size": 131072, 00:27:30.289 "runtime": 1.23109, 00:27:30.289 "iops": 18913.320715788446, 00:27:30.289 "mibps": 2364.165089473556, 00:27:30.289 "io_failed": 0, 00:27:30.289 "io_timeout": 0, 00:27:30.289 "avg_latency_us": 49.629076685211366, 00:27:30.289 "min_latency_us": 28.947692307692307, 00:27:30.289 "max_latency_us": 1663.6061538461538 00:27:30.289 } 00:27:30.289 ], 00:27:30.289 "core_count": 1 00:27:30.289 } 00:27:30.289 13:47:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:30.289 13:47:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 62206 00:27:30.289 13:47:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 62206 ']' 00:27:30.289 13:47:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 62206 00:27:30.289 13:47:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:27:30.289 13:47:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:30.289 13:47:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 62206 00:27:30.289 13:47:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:27:30.289 13:47:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:27:30.289 13:47:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 62206' 00:27:30.289 killing process with pid 62206 00:27:30.289 13:47:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 62206 00:27:30.289 [2024-11-20 13:47:37.216033] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:27:30.289 13:47:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 62206 00:27:30.289 [2024-11-20 13:47:37.303945] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:27:31.249 13:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:27:31.249 13:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.Khi4dcrfi8 00:27:31.249 13:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:27:31.249 13:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:27:31.249 13:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:27:31.249 13:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:27:31.249 13:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:27:31.249 13:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:27:31.249 00:27:31.249 real 0m3.608s 00:27:31.249 user 0m4.288s 00:27:31.249 sys 0m0.408s 00:27:31.249 ************************************ 00:27:31.249 END TEST raid_write_error_test 00:27:31.249 ************************************ 00:27:31.249 13:47:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:27:31.249 13:47:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:31.249 13:47:38 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:27:31.249 13:47:38 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:27:31.249 13:47:38 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 3 false 00:27:31.249 13:47:38 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:27:31.249 13:47:38 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:27:31.249 13:47:38 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:27:31.249 ************************************ 00:27:31.249 START TEST raid_state_function_test 00:27:31.249 ************************************ 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 3 false 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:27:31.249 Process raid pid: 62347 00:27:31.249 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=62347 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 62347' 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 62347 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 62347 ']' 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:31.249 13:47:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:27:31.249 [2024-11-20 13:47:38.230721] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:27:31.250 [2024-11-20 13:47:38.230870] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:27:31.507 [2024-11-20 13:47:38.395044] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:31.507 [2024-11-20 13:47:38.513798] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:27:31.765 [2024-11-20 13:47:38.663387] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:31.765 [2024-11-20 13:47:38.663470] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:32.342 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:32.342 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:27:32.342 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:27:32.342 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:32.342 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:32.342 [2024-11-20 13:47:39.097017] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:27:32.342 [2024-11-20 13:47:39.097074] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:27:32.342 [2024-11-20 13:47:39.097085] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:27:32.342 [2024-11-20 13:47:39.097095] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:27:32.342 [2024-11-20 13:47:39.097102] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:27:32.342 [2024-11-20 13:47:39.097111] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:27:32.342 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:32.342 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:27:32.342 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:32.342 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:32.342 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:32.342 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:32.342 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:32.342 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:32.342 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:32.342 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:32.342 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:32.342 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:32.342 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:32.343 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:32.343 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:32.343 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:32.343 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:32.343 "name": "Existed_Raid", 00:27:32.343 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:32.343 "strip_size_kb": 64, 00:27:32.343 "state": "configuring", 00:27:32.343 "raid_level": "raid0", 00:27:32.343 "superblock": false, 00:27:32.343 "num_base_bdevs": 3, 00:27:32.343 "num_base_bdevs_discovered": 0, 00:27:32.343 "num_base_bdevs_operational": 3, 00:27:32.343 "base_bdevs_list": [ 00:27:32.343 { 00:27:32.343 "name": "BaseBdev1", 00:27:32.343 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:32.343 "is_configured": false, 00:27:32.343 "data_offset": 0, 00:27:32.343 "data_size": 0 00:27:32.343 }, 00:27:32.343 { 00:27:32.343 "name": "BaseBdev2", 00:27:32.343 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:32.343 "is_configured": false, 00:27:32.343 "data_offset": 0, 00:27:32.343 "data_size": 0 00:27:32.343 }, 00:27:32.343 { 00:27:32.343 "name": "BaseBdev3", 00:27:32.343 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:32.343 "is_configured": false, 00:27:32.343 "data_offset": 0, 00:27:32.343 "data_size": 0 00:27:32.343 } 00:27:32.343 ] 00:27:32.343 }' 00:27:32.343 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:32.343 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:32.602 [2024-11-20 13:47:39.437040] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:27:32.602 [2024-11-20 13:47:39.437078] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:32.602 [2024-11-20 13:47:39.445045] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:27:32.602 [2024-11-20 13:47:39.445087] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:27:32.602 [2024-11-20 13:47:39.445095] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:27:32.602 [2024-11-20 13:47:39.445105] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:27:32.602 [2024-11-20 13:47:39.445111] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:27:32.602 [2024-11-20 13:47:39.445120] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:32.602 [2024-11-20 13:47:39.479795] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:27:32.602 BaseBdev1 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:32.602 [ 00:27:32.602 { 00:27:32.602 "name": "BaseBdev1", 00:27:32.602 "aliases": [ 00:27:32.602 "cc284ecd-b07b-45a3-9259-7286032130eb" 00:27:32.602 ], 00:27:32.602 "product_name": "Malloc disk", 00:27:32.602 "block_size": 512, 00:27:32.602 "num_blocks": 65536, 00:27:32.602 "uuid": "cc284ecd-b07b-45a3-9259-7286032130eb", 00:27:32.602 "assigned_rate_limits": { 00:27:32.602 "rw_ios_per_sec": 0, 00:27:32.602 "rw_mbytes_per_sec": 0, 00:27:32.602 "r_mbytes_per_sec": 0, 00:27:32.602 "w_mbytes_per_sec": 0 00:27:32.602 }, 00:27:32.602 "claimed": true, 00:27:32.602 "claim_type": "exclusive_write", 00:27:32.602 "zoned": false, 00:27:32.602 "supported_io_types": { 00:27:32.602 "read": true, 00:27:32.602 "write": true, 00:27:32.602 "unmap": true, 00:27:32.602 "flush": true, 00:27:32.602 "reset": true, 00:27:32.602 "nvme_admin": false, 00:27:32.602 "nvme_io": false, 00:27:32.602 "nvme_io_md": false, 00:27:32.602 "write_zeroes": true, 00:27:32.602 "zcopy": true, 00:27:32.602 "get_zone_info": false, 00:27:32.602 "zone_management": false, 00:27:32.602 "zone_append": false, 00:27:32.602 "compare": false, 00:27:32.602 "compare_and_write": false, 00:27:32.602 "abort": true, 00:27:32.602 "seek_hole": false, 00:27:32.602 "seek_data": false, 00:27:32.602 "copy": true, 00:27:32.602 "nvme_iov_md": false 00:27:32.602 }, 00:27:32.602 "memory_domains": [ 00:27:32.602 { 00:27:32.602 "dma_device_id": "system", 00:27:32.602 "dma_device_type": 1 00:27:32.602 }, 00:27:32.602 { 00:27:32.602 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:32.602 "dma_device_type": 2 00:27:32.602 } 00:27:32.602 ], 00:27:32.602 "driver_specific": {} 00:27:32.602 } 00:27:32.602 ] 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:32.602 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:32.603 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:32.603 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:32.603 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:32.603 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:32.603 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:32.603 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:32.603 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:32.603 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:32.603 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:32.603 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:32.603 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:32.603 "name": "Existed_Raid", 00:27:32.603 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:32.603 "strip_size_kb": 64, 00:27:32.603 "state": "configuring", 00:27:32.603 "raid_level": "raid0", 00:27:32.603 "superblock": false, 00:27:32.603 "num_base_bdevs": 3, 00:27:32.603 "num_base_bdevs_discovered": 1, 00:27:32.603 "num_base_bdevs_operational": 3, 00:27:32.603 "base_bdevs_list": [ 00:27:32.603 { 00:27:32.603 "name": "BaseBdev1", 00:27:32.603 "uuid": "cc284ecd-b07b-45a3-9259-7286032130eb", 00:27:32.603 "is_configured": true, 00:27:32.603 "data_offset": 0, 00:27:32.603 "data_size": 65536 00:27:32.603 }, 00:27:32.603 { 00:27:32.603 "name": "BaseBdev2", 00:27:32.603 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:32.603 "is_configured": false, 00:27:32.603 "data_offset": 0, 00:27:32.603 "data_size": 0 00:27:32.603 }, 00:27:32.603 { 00:27:32.603 "name": "BaseBdev3", 00:27:32.603 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:32.603 "is_configured": false, 00:27:32.603 "data_offset": 0, 00:27:32.603 "data_size": 0 00:27:32.603 } 00:27:32.603 ] 00:27:32.603 }' 00:27:32.603 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:32.603 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:32.861 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:27:32.861 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:32.861 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:32.861 [2024-11-20 13:47:39.811917] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:27:32.861 [2024-11-20 13:47:39.812088] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:27:32.861 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:32.861 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:27:32.861 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:32.861 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:32.861 [2024-11-20 13:47:39.819988] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:27:32.861 [2024-11-20 13:47:39.822052] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:27:32.861 [2024-11-20 13:47:39.822173] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:27:32.861 [2024-11-20 13:47:39.822232] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:27:32.861 [2024-11-20 13:47:39.822259] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:27:32.861 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:32.861 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:27:32.861 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:27:32.861 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:27:32.861 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:32.861 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:32.861 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:32.861 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:32.861 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:32.861 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:32.861 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:32.861 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:32.861 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:32.861 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:32.861 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:32.861 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:32.861 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:32.861 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:32.861 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:32.861 "name": "Existed_Raid", 00:27:32.861 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:32.861 "strip_size_kb": 64, 00:27:32.861 "state": "configuring", 00:27:32.861 "raid_level": "raid0", 00:27:32.861 "superblock": false, 00:27:32.861 "num_base_bdevs": 3, 00:27:32.861 "num_base_bdevs_discovered": 1, 00:27:32.861 "num_base_bdevs_operational": 3, 00:27:32.861 "base_bdevs_list": [ 00:27:32.861 { 00:27:32.861 "name": "BaseBdev1", 00:27:32.861 "uuid": "cc284ecd-b07b-45a3-9259-7286032130eb", 00:27:32.861 "is_configured": true, 00:27:32.861 "data_offset": 0, 00:27:32.861 "data_size": 65536 00:27:32.861 }, 00:27:32.861 { 00:27:32.861 "name": "BaseBdev2", 00:27:32.861 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:32.861 "is_configured": false, 00:27:32.861 "data_offset": 0, 00:27:32.861 "data_size": 0 00:27:32.861 }, 00:27:32.861 { 00:27:32.861 "name": "BaseBdev3", 00:27:32.861 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:32.861 "is_configured": false, 00:27:32.861 "data_offset": 0, 00:27:32.861 "data_size": 0 00:27:32.861 } 00:27:32.861 ] 00:27:32.861 }' 00:27:32.862 13:47:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:32.862 13:47:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:33.119 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:27:33.119 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:33.119 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:33.119 [2024-11-20 13:47:40.164721] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:27:33.119 BaseBdev2 00:27:33.119 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:33.119 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:27:33.119 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:27:33.119 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:27:33.119 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:27:33.119 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:27:33.119 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:27:33.119 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:27:33.119 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:33.119 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:33.119 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:33.119 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:27:33.119 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:33.119 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:33.376 [ 00:27:33.376 { 00:27:33.376 "name": "BaseBdev2", 00:27:33.376 "aliases": [ 00:27:33.376 "13b8571a-958b-4503-8b3e-96007cd44fe5" 00:27:33.376 ], 00:27:33.376 "product_name": "Malloc disk", 00:27:33.376 "block_size": 512, 00:27:33.376 "num_blocks": 65536, 00:27:33.376 "uuid": "13b8571a-958b-4503-8b3e-96007cd44fe5", 00:27:33.376 "assigned_rate_limits": { 00:27:33.376 "rw_ios_per_sec": 0, 00:27:33.376 "rw_mbytes_per_sec": 0, 00:27:33.376 "r_mbytes_per_sec": 0, 00:27:33.376 "w_mbytes_per_sec": 0 00:27:33.376 }, 00:27:33.376 "claimed": true, 00:27:33.376 "claim_type": "exclusive_write", 00:27:33.376 "zoned": false, 00:27:33.376 "supported_io_types": { 00:27:33.376 "read": true, 00:27:33.376 "write": true, 00:27:33.376 "unmap": true, 00:27:33.376 "flush": true, 00:27:33.376 "reset": true, 00:27:33.376 "nvme_admin": false, 00:27:33.376 "nvme_io": false, 00:27:33.376 "nvme_io_md": false, 00:27:33.376 "write_zeroes": true, 00:27:33.376 "zcopy": true, 00:27:33.376 "get_zone_info": false, 00:27:33.376 "zone_management": false, 00:27:33.376 "zone_append": false, 00:27:33.376 "compare": false, 00:27:33.376 "compare_and_write": false, 00:27:33.376 "abort": true, 00:27:33.376 "seek_hole": false, 00:27:33.376 "seek_data": false, 00:27:33.376 "copy": true, 00:27:33.376 "nvme_iov_md": false 00:27:33.376 }, 00:27:33.376 "memory_domains": [ 00:27:33.376 { 00:27:33.376 "dma_device_id": "system", 00:27:33.376 "dma_device_type": 1 00:27:33.376 }, 00:27:33.376 { 00:27:33.376 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:33.376 "dma_device_type": 2 00:27:33.376 } 00:27:33.376 ], 00:27:33.376 "driver_specific": {} 00:27:33.376 } 00:27:33.376 ] 00:27:33.376 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:33.376 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:27:33.376 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:27:33.376 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:27:33.376 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:27:33.376 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:33.376 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:33.376 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:33.376 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:33.376 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:33.376 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:33.376 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:33.377 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:33.377 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:33.377 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:33.377 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:33.377 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:33.377 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:33.377 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:33.377 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:33.377 "name": "Existed_Raid", 00:27:33.377 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:33.377 "strip_size_kb": 64, 00:27:33.377 "state": "configuring", 00:27:33.377 "raid_level": "raid0", 00:27:33.377 "superblock": false, 00:27:33.377 "num_base_bdevs": 3, 00:27:33.377 "num_base_bdevs_discovered": 2, 00:27:33.377 "num_base_bdevs_operational": 3, 00:27:33.377 "base_bdevs_list": [ 00:27:33.377 { 00:27:33.377 "name": "BaseBdev1", 00:27:33.377 "uuid": "cc284ecd-b07b-45a3-9259-7286032130eb", 00:27:33.377 "is_configured": true, 00:27:33.377 "data_offset": 0, 00:27:33.377 "data_size": 65536 00:27:33.377 }, 00:27:33.377 { 00:27:33.377 "name": "BaseBdev2", 00:27:33.377 "uuid": "13b8571a-958b-4503-8b3e-96007cd44fe5", 00:27:33.377 "is_configured": true, 00:27:33.377 "data_offset": 0, 00:27:33.377 "data_size": 65536 00:27:33.377 }, 00:27:33.377 { 00:27:33.377 "name": "BaseBdev3", 00:27:33.377 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:33.377 "is_configured": false, 00:27:33.377 "data_offset": 0, 00:27:33.377 "data_size": 0 00:27:33.377 } 00:27:33.377 ] 00:27:33.377 }' 00:27:33.377 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:33.377 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:33.634 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:27:33.634 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:33.634 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:33.634 [2024-11-20 13:47:40.556892] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:27:33.634 [2024-11-20 13:47:40.557192] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:27:33.634 [2024-11-20 13:47:40.557234] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:27:33.634 [2024-11-20 13:47:40.557586] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:27:33.634 [2024-11-20 13:47:40.557837] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:27:33.634 [2024-11-20 13:47:40.557918] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:27:33.634 [2024-11-20 13:47:40.558268] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:33.634 BaseBdev3 00:27:33.634 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:33.634 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:27:33.634 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:27:33.634 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:27:33.634 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:33.635 [ 00:27:33.635 { 00:27:33.635 "name": "BaseBdev3", 00:27:33.635 "aliases": [ 00:27:33.635 "d769478f-f955-4dd4-8e91-f36270d5dbf4" 00:27:33.635 ], 00:27:33.635 "product_name": "Malloc disk", 00:27:33.635 "block_size": 512, 00:27:33.635 "num_blocks": 65536, 00:27:33.635 "uuid": "d769478f-f955-4dd4-8e91-f36270d5dbf4", 00:27:33.635 "assigned_rate_limits": { 00:27:33.635 "rw_ios_per_sec": 0, 00:27:33.635 "rw_mbytes_per_sec": 0, 00:27:33.635 "r_mbytes_per_sec": 0, 00:27:33.635 "w_mbytes_per_sec": 0 00:27:33.635 }, 00:27:33.635 "claimed": true, 00:27:33.635 "claim_type": "exclusive_write", 00:27:33.635 "zoned": false, 00:27:33.635 "supported_io_types": { 00:27:33.635 "read": true, 00:27:33.635 "write": true, 00:27:33.635 "unmap": true, 00:27:33.635 "flush": true, 00:27:33.635 "reset": true, 00:27:33.635 "nvme_admin": false, 00:27:33.635 "nvme_io": false, 00:27:33.635 "nvme_io_md": false, 00:27:33.635 "write_zeroes": true, 00:27:33.635 "zcopy": true, 00:27:33.635 "get_zone_info": false, 00:27:33.635 "zone_management": false, 00:27:33.635 "zone_append": false, 00:27:33.635 "compare": false, 00:27:33.635 "compare_and_write": false, 00:27:33.635 "abort": true, 00:27:33.635 "seek_hole": false, 00:27:33.635 "seek_data": false, 00:27:33.635 "copy": true, 00:27:33.635 "nvme_iov_md": false 00:27:33.635 }, 00:27:33.635 "memory_domains": [ 00:27:33.635 { 00:27:33.635 "dma_device_id": "system", 00:27:33.635 "dma_device_type": 1 00:27:33.635 }, 00:27:33.635 { 00:27:33.635 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:33.635 "dma_device_type": 2 00:27:33.635 } 00:27:33.635 ], 00:27:33.635 "driver_specific": {} 00:27:33.635 } 00:27:33.635 ] 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:33.635 "name": "Existed_Raid", 00:27:33.635 "uuid": "4f2719b1-af3c-4321-b4ec-f38d66618ddd", 00:27:33.635 "strip_size_kb": 64, 00:27:33.635 "state": "online", 00:27:33.635 "raid_level": "raid0", 00:27:33.635 "superblock": false, 00:27:33.635 "num_base_bdevs": 3, 00:27:33.635 "num_base_bdevs_discovered": 3, 00:27:33.635 "num_base_bdevs_operational": 3, 00:27:33.635 "base_bdevs_list": [ 00:27:33.635 { 00:27:33.635 "name": "BaseBdev1", 00:27:33.635 "uuid": "cc284ecd-b07b-45a3-9259-7286032130eb", 00:27:33.635 "is_configured": true, 00:27:33.635 "data_offset": 0, 00:27:33.635 "data_size": 65536 00:27:33.635 }, 00:27:33.635 { 00:27:33.635 "name": "BaseBdev2", 00:27:33.635 "uuid": "13b8571a-958b-4503-8b3e-96007cd44fe5", 00:27:33.635 "is_configured": true, 00:27:33.635 "data_offset": 0, 00:27:33.635 "data_size": 65536 00:27:33.635 }, 00:27:33.635 { 00:27:33.635 "name": "BaseBdev3", 00:27:33.635 "uuid": "d769478f-f955-4dd4-8e91-f36270d5dbf4", 00:27:33.635 "is_configured": true, 00:27:33.635 "data_offset": 0, 00:27:33.635 "data_size": 65536 00:27:33.635 } 00:27:33.635 ] 00:27:33.635 }' 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:33.635 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:33.893 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:27:33.893 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:27:33.893 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:27:33.893 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:27:33.893 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:27:33.893 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:27:33.893 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:27:33.893 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:27:33.893 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:33.893 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:33.893 [2024-11-20 13:47:40.889359] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:33.893 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:33.893 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:27:33.893 "name": "Existed_Raid", 00:27:33.893 "aliases": [ 00:27:33.893 "4f2719b1-af3c-4321-b4ec-f38d66618ddd" 00:27:33.893 ], 00:27:33.893 "product_name": "Raid Volume", 00:27:33.893 "block_size": 512, 00:27:33.893 "num_blocks": 196608, 00:27:33.893 "uuid": "4f2719b1-af3c-4321-b4ec-f38d66618ddd", 00:27:33.893 "assigned_rate_limits": { 00:27:33.893 "rw_ios_per_sec": 0, 00:27:33.893 "rw_mbytes_per_sec": 0, 00:27:33.893 "r_mbytes_per_sec": 0, 00:27:33.893 "w_mbytes_per_sec": 0 00:27:33.893 }, 00:27:33.893 "claimed": false, 00:27:33.893 "zoned": false, 00:27:33.893 "supported_io_types": { 00:27:33.893 "read": true, 00:27:33.893 "write": true, 00:27:33.893 "unmap": true, 00:27:33.893 "flush": true, 00:27:33.893 "reset": true, 00:27:33.893 "nvme_admin": false, 00:27:33.893 "nvme_io": false, 00:27:33.893 "nvme_io_md": false, 00:27:33.893 "write_zeroes": true, 00:27:33.893 "zcopy": false, 00:27:33.893 "get_zone_info": false, 00:27:33.893 "zone_management": false, 00:27:33.893 "zone_append": false, 00:27:33.893 "compare": false, 00:27:33.893 "compare_and_write": false, 00:27:33.893 "abort": false, 00:27:33.893 "seek_hole": false, 00:27:33.893 "seek_data": false, 00:27:33.893 "copy": false, 00:27:33.893 "nvme_iov_md": false 00:27:33.893 }, 00:27:33.893 "memory_domains": [ 00:27:33.893 { 00:27:33.893 "dma_device_id": "system", 00:27:33.893 "dma_device_type": 1 00:27:33.893 }, 00:27:33.893 { 00:27:33.893 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:33.893 "dma_device_type": 2 00:27:33.893 }, 00:27:33.893 { 00:27:33.893 "dma_device_id": "system", 00:27:33.893 "dma_device_type": 1 00:27:33.893 }, 00:27:33.893 { 00:27:33.893 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:33.893 "dma_device_type": 2 00:27:33.893 }, 00:27:33.893 { 00:27:33.893 "dma_device_id": "system", 00:27:33.893 "dma_device_type": 1 00:27:33.893 }, 00:27:33.893 { 00:27:33.893 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:33.893 "dma_device_type": 2 00:27:33.893 } 00:27:33.893 ], 00:27:33.893 "driver_specific": { 00:27:33.893 "raid": { 00:27:33.893 "uuid": "4f2719b1-af3c-4321-b4ec-f38d66618ddd", 00:27:33.893 "strip_size_kb": 64, 00:27:33.893 "state": "online", 00:27:33.893 "raid_level": "raid0", 00:27:33.893 "superblock": false, 00:27:33.893 "num_base_bdevs": 3, 00:27:33.893 "num_base_bdevs_discovered": 3, 00:27:33.893 "num_base_bdevs_operational": 3, 00:27:33.893 "base_bdevs_list": [ 00:27:33.893 { 00:27:33.893 "name": "BaseBdev1", 00:27:33.893 "uuid": "cc284ecd-b07b-45a3-9259-7286032130eb", 00:27:33.893 "is_configured": true, 00:27:33.893 "data_offset": 0, 00:27:33.893 "data_size": 65536 00:27:33.893 }, 00:27:33.893 { 00:27:33.893 "name": "BaseBdev2", 00:27:33.893 "uuid": "13b8571a-958b-4503-8b3e-96007cd44fe5", 00:27:33.893 "is_configured": true, 00:27:33.893 "data_offset": 0, 00:27:33.893 "data_size": 65536 00:27:33.893 }, 00:27:33.893 { 00:27:33.893 "name": "BaseBdev3", 00:27:33.893 "uuid": "d769478f-f955-4dd4-8e91-f36270d5dbf4", 00:27:33.893 "is_configured": true, 00:27:33.893 "data_offset": 0, 00:27:33.893 "data_size": 65536 00:27:33.893 } 00:27:33.893 ] 00:27:33.893 } 00:27:33.893 } 00:27:33.893 }' 00:27:33.893 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:27:34.151 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:27:34.151 BaseBdev2 00:27:34.151 BaseBdev3' 00:27:34.151 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:34.151 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:27:34.151 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:34.151 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:27:34.151 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:34.151 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:34.151 13:47:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:34.151 13:47:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:34.151 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:34.151 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:34.151 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:34.151 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:27:34.151 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:34.151 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:34.151 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:34.151 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:34.151 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:34.151 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:34.151 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:34.151 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:27:34.151 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:34.151 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:34.151 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:34.151 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:34.151 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:34.151 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:34.151 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:27:34.151 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:34.151 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:34.151 [2024-11-20 13:47:41.113130] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:27:34.151 [2024-11-20 13:47:41.113159] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:27:34.152 [2024-11-20 13:47:41.113219] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:34.152 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:34.152 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:27:34.152 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:27:34.152 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:27:34.152 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:27:34.152 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:27:34.152 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 2 00:27:34.152 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:34.152 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:27:34.152 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:34.152 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:34.152 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:34.152 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:34.152 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:34.152 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:34.152 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:34.152 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:34.152 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:34.152 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:34.152 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:34.152 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:34.409 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:34.409 "name": "Existed_Raid", 00:27:34.409 "uuid": "4f2719b1-af3c-4321-b4ec-f38d66618ddd", 00:27:34.409 "strip_size_kb": 64, 00:27:34.409 "state": "offline", 00:27:34.409 "raid_level": "raid0", 00:27:34.409 "superblock": false, 00:27:34.409 "num_base_bdevs": 3, 00:27:34.409 "num_base_bdevs_discovered": 2, 00:27:34.410 "num_base_bdevs_operational": 2, 00:27:34.410 "base_bdevs_list": [ 00:27:34.410 { 00:27:34.410 "name": null, 00:27:34.410 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:34.410 "is_configured": false, 00:27:34.410 "data_offset": 0, 00:27:34.410 "data_size": 65536 00:27:34.410 }, 00:27:34.410 { 00:27:34.410 "name": "BaseBdev2", 00:27:34.410 "uuid": "13b8571a-958b-4503-8b3e-96007cd44fe5", 00:27:34.410 "is_configured": true, 00:27:34.410 "data_offset": 0, 00:27:34.410 "data_size": 65536 00:27:34.410 }, 00:27:34.410 { 00:27:34.410 "name": "BaseBdev3", 00:27:34.410 "uuid": "d769478f-f955-4dd4-8e91-f36270d5dbf4", 00:27:34.410 "is_configured": true, 00:27:34.410 "data_offset": 0, 00:27:34.410 "data_size": 65536 00:27:34.410 } 00:27:34.410 ] 00:27:34.410 }' 00:27:34.410 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:34.410 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:34.668 [2024-11-20 13:47:41.524692] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:34.668 [2024-11-20 13:47:41.627344] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:27:34.668 [2024-11-20 13:47:41.627576] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:34.668 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:34.926 BaseBdev2 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:34.926 [ 00:27:34.926 { 00:27:34.926 "name": "BaseBdev2", 00:27:34.926 "aliases": [ 00:27:34.926 "7dcdb139-7abd-4b2e-9f24-4c0e75b17ffc" 00:27:34.926 ], 00:27:34.926 "product_name": "Malloc disk", 00:27:34.926 "block_size": 512, 00:27:34.926 "num_blocks": 65536, 00:27:34.926 "uuid": "7dcdb139-7abd-4b2e-9f24-4c0e75b17ffc", 00:27:34.926 "assigned_rate_limits": { 00:27:34.926 "rw_ios_per_sec": 0, 00:27:34.926 "rw_mbytes_per_sec": 0, 00:27:34.926 "r_mbytes_per_sec": 0, 00:27:34.926 "w_mbytes_per_sec": 0 00:27:34.926 }, 00:27:34.926 "claimed": false, 00:27:34.926 "zoned": false, 00:27:34.926 "supported_io_types": { 00:27:34.926 "read": true, 00:27:34.926 "write": true, 00:27:34.926 "unmap": true, 00:27:34.926 "flush": true, 00:27:34.926 "reset": true, 00:27:34.926 "nvme_admin": false, 00:27:34.926 "nvme_io": false, 00:27:34.926 "nvme_io_md": false, 00:27:34.926 "write_zeroes": true, 00:27:34.926 "zcopy": true, 00:27:34.926 "get_zone_info": false, 00:27:34.926 "zone_management": false, 00:27:34.926 "zone_append": false, 00:27:34.926 "compare": false, 00:27:34.926 "compare_and_write": false, 00:27:34.926 "abort": true, 00:27:34.926 "seek_hole": false, 00:27:34.926 "seek_data": false, 00:27:34.926 "copy": true, 00:27:34.926 "nvme_iov_md": false 00:27:34.926 }, 00:27:34.926 "memory_domains": [ 00:27:34.926 { 00:27:34.926 "dma_device_id": "system", 00:27:34.926 "dma_device_type": 1 00:27:34.926 }, 00:27:34.926 { 00:27:34.926 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:34.926 "dma_device_type": 2 00:27:34.926 } 00:27:34.926 ], 00:27:34.926 "driver_specific": {} 00:27:34.926 } 00:27:34.926 ] 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:34.926 BaseBdev3 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:34.926 [ 00:27:34.926 { 00:27:34.926 "name": "BaseBdev3", 00:27:34.926 "aliases": [ 00:27:34.926 "9b4e0f65-0c94-4839-bc75-a9b125b75b7e" 00:27:34.926 ], 00:27:34.926 "product_name": "Malloc disk", 00:27:34.926 "block_size": 512, 00:27:34.926 "num_blocks": 65536, 00:27:34.926 "uuid": "9b4e0f65-0c94-4839-bc75-a9b125b75b7e", 00:27:34.926 "assigned_rate_limits": { 00:27:34.926 "rw_ios_per_sec": 0, 00:27:34.926 "rw_mbytes_per_sec": 0, 00:27:34.926 "r_mbytes_per_sec": 0, 00:27:34.926 "w_mbytes_per_sec": 0 00:27:34.926 }, 00:27:34.926 "claimed": false, 00:27:34.926 "zoned": false, 00:27:34.926 "supported_io_types": { 00:27:34.926 "read": true, 00:27:34.926 "write": true, 00:27:34.926 "unmap": true, 00:27:34.926 "flush": true, 00:27:34.926 "reset": true, 00:27:34.926 "nvme_admin": false, 00:27:34.926 "nvme_io": false, 00:27:34.926 "nvme_io_md": false, 00:27:34.926 "write_zeroes": true, 00:27:34.926 "zcopy": true, 00:27:34.926 "get_zone_info": false, 00:27:34.926 "zone_management": false, 00:27:34.926 "zone_append": false, 00:27:34.926 "compare": false, 00:27:34.926 "compare_and_write": false, 00:27:34.926 "abort": true, 00:27:34.926 "seek_hole": false, 00:27:34.926 "seek_data": false, 00:27:34.926 "copy": true, 00:27:34.926 "nvme_iov_md": false 00:27:34.926 }, 00:27:34.926 "memory_domains": [ 00:27:34.926 { 00:27:34.926 "dma_device_id": "system", 00:27:34.926 "dma_device_type": 1 00:27:34.926 }, 00:27:34.926 { 00:27:34.926 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:34.926 "dma_device_type": 2 00:27:34.926 } 00:27:34.926 ], 00:27:34.926 "driver_specific": {} 00:27:34.926 } 00:27:34.926 ] 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:34.926 [2024-11-20 13:47:41.852108] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:27:34.926 [2024-11-20 13:47:41.852273] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:27:34.926 [2024-11-20 13:47:41.852351] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:27:34.926 [2024-11-20 13:47:41.854363] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:34.926 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:34.927 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:34.927 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:34.927 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:34.927 "name": "Existed_Raid", 00:27:34.927 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:34.927 "strip_size_kb": 64, 00:27:34.927 "state": "configuring", 00:27:34.927 "raid_level": "raid0", 00:27:34.927 "superblock": false, 00:27:34.927 "num_base_bdevs": 3, 00:27:34.927 "num_base_bdevs_discovered": 2, 00:27:34.927 "num_base_bdevs_operational": 3, 00:27:34.927 "base_bdevs_list": [ 00:27:34.927 { 00:27:34.927 "name": "BaseBdev1", 00:27:34.927 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:34.927 "is_configured": false, 00:27:34.927 "data_offset": 0, 00:27:34.927 "data_size": 0 00:27:34.927 }, 00:27:34.927 { 00:27:34.927 "name": "BaseBdev2", 00:27:34.927 "uuid": "7dcdb139-7abd-4b2e-9f24-4c0e75b17ffc", 00:27:34.927 "is_configured": true, 00:27:34.927 "data_offset": 0, 00:27:34.927 "data_size": 65536 00:27:34.927 }, 00:27:34.927 { 00:27:34.927 "name": "BaseBdev3", 00:27:34.927 "uuid": "9b4e0f65-0c94-4839-bc75-a9b125b75b7e", 00:27:34.927 "is_configured": true, 00:27:34.927 "data_offset": 0, 00:27:34.927 "data_size": 65536 00:27:34.927 } 00:27:34.927 ] 00:27:34.927 }' 00:27:34.927 13:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:34.927 13:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:35.184 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:27:35.184 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:35.184 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:35.184 [2024-11-20 13:47:42.160190] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:27:35.184 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:35.184 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:27:35.184 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:35.184 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:35.184 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:35.184 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:35.184 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:35.184 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:35.184 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:35.184 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:35.184 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:35.184 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:35.184 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:35.184 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:35.184 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:35.185 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:35.185 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:35.185 "name": "Existed_Raid", 00:27:35.185 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:35.185 "strip_size_kb": 64, 00:27:35.185 "state": "configuring", 00:27:35.185 "raid_level": "raid0", 00:27:35.185 "superblock": false, 00:27:35.185 "num_base_bdevs": 3, 00:27:35.185 "num_base_bdevs_discovered": 1, 00:27:35.185 "num_base_bdevs_operational": 3, 00:27:35.185 "base_bdevs_list": [ 00:27:35.185 { 00:27:35.185 "name": "BaseBdev1", 00:27:35.185 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:35.185 "is_configured": false, 00:27:35.185 "data_offset": 0, 00:27:35.185 "data_size": 0 00:27:35.185 }, 00:27:35.185 { 00:27:35.185 "name": null, 00:27:35.185 "uuid": "7dcdb139-7abd-4b2e-9f24-4c0e75b17ffc", 00:27:35.185 "is_configured": false, 00:27:35.185 "data_offset": 0, 00:27:35.185 "data_size": 65536 00:27:35.185 }, 00:27:35.185 { 00:27:35.185 "name": "BaseBdev3", 00:27:35.185 "uuid": "9b4e0f65-0c94-4839-bc75-a9b125b75b7e", 00:27:35.185 "is_configured": true, 00:27:35.185 "data_offset": 0, 00:27:35.185 "data_size": 65536 00:27:35.185 } 00:27:35.185 ] 00:27:35.185 }' 00:27:35.185 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:35.185 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:35.442 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:35.442 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:35.442 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:35.442 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:27:35.442 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:35.699 [2024-11-20 13:47:42.533252] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:27:35.699 BaseBdev1 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:35.699 [ 00:27:35.699 { 00:27:35.699 "name": "BaseBdev1", 00:27:35.699 "aliases": [ 00:27:35.699 "922d6536-121c-41de-8d3e-59e21dc78cf3" 00:27:35.699 ], 00:27:35.699 "product_name": "Malloc disk", 00:27:35.699 "block_size": 512, 00:27:35.699 "num_blocks": 65536, 00:27:35.699 "uuid": "922d6536-121c-41de-8d3e-59e21dc78cf3", 00:27:35.699 "assigned_rate_limits": { 00:27:35.699 "rw_ios_per_sec": 0, 00:27:35.699 "rw_mbytes_per_sec": 0, 00:27:35.699 "r_mbytes_per_sec": 0, 00:27:35.699 "w_mbytes_per_sec": 0 00:27:35.699 }, 00:27:35.699 "claimed": true, 00:27:35.699 "claim_type": "exclusive_write", 00:27:35.699 "zoned": false, 00:27:35.699 "supported_io_types": { 00:27:35.699 "read": true, 00:27:35.699 "write": true, 00:27:35.699 "unmap": true, 00:27:35.699 "flush": true, 00:27:35.699 "reset": true, 00:27:35.699 "nvme_admin": false, 00:27:35.699 "nvme_io": false, 00:27:35.699 "nvme_io_md": false, 00:27:35.699 "write_zeroes": true, 00:27:35.699 "zcopy": true, 00:27:35.699 "get_zone_info": false, 00:27:35.699 "zone_management": false, 00:27:35.699 "zone_append": false, 00:27:35.699 "compare": false, 00:27:35.699 "compare_and_write": false, 00:27:35.699 "abort": true, 00:27:35.699 "seek_hole": false, 00:27:35.699 "seek_data": false, 00:27:35.699 "copy": true, 00:27:35.699 "nvme_iov_md": false 00:27:35.699 }, 00:27:35.699 "memory_domains": [ 00:27:35.699 { 00:27:35.699 "dma_device_id": "system", 00:27:35.699 "dma_device_type": 1 00:27:35.699 }, 00:27:35.699 { 00:27:35.699 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:35.699 "dma_device_type": 2 00:27:35.699 } 00:27:35.699 ], 00:27:35.699 "driver_specific": {} 00:27:35.699 } 00:27:35.699 ] 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:35.699 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:35.700 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:35.700 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:35.700 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:35.700 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:35.700 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:35.700 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:35.700 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:35.700 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:35.700 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:35.700 "name": "Existed_Raid", 00:27:35.700 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:35.700 "strip_size_kb": 64, 00:27:35.700 "state": "configuring", 00:27:35.700 "raid_level": "raid0", 00:27:35.700 "superblock": false, 00:27:35.700 "num_base_bdevs": 3, 00:27:35.700 "num_base_bdevs_discovered": 2, 00:27:35.700 "num_base_bdevs_operational": 3, 00:27:35.700 "base_bdevs_list": [ 00:27:35.700 { 00:27:35.700 "name": "BaseBdev1", 00:27:35.700 "uuid": "922d6536-121c-41de-8d3e-59e21dc78cf3", 00:27:35.700 "is_configured": true, 00:27:35.700 "data_offset": 0, 00:27:35.700 "data_size": 65536 00:27:35.700 }, 00:27:35.700 { 00:27:35.700 "name": null, 00:27:35.700 "uuid": "7dcdb139-7abd-4b2e-9f24-4c0e75b17ffc", 00:27:35.700 "is_configured": false, 00:27:35.700 "data_offset": 0, 00:27:35.700 "data_size": 65536 00:27:35.700 }, 00:27:35.700 { 00:27:35.700 "name": "BaseBdev3", 00:27:35.700 "uuid": "9b4e0f65-0c94-4839-bc75-a9b125b75b7e", 00:27:35.700 "is_configured": true, 00:27:35.700 "data_offset": 0, 00:27:35.700 "data_size": 65536 00:27:35.700 } 00:27:35.700 ] 00:27:35.700 }' 00:27:35.700 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:35.700 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:35.957 [2024-11-20 13:47:42.901365] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:35.957 "name": "Existed_Raid", 00:27:35.957 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:35.957 "strip_size_kb": 64, 00:27:35.957 "state": "configuring", 00:27:35.957 "raid_level": "raid0", 00:27:35.957 "superblock": false, 00:27:35.957 "num_base_bdevs": 3, 00:27:35.957 "num_base_bdevs_discovered": 1, 00:27:35.957 "num_base_bdevs_operational": 3, 00:27:35.957 "base_bdevs_list": [ 00:27:35.957 { 00:27:35.957 "name": "BaseBdev1", 00:27:35.957 "uuid": "922d6536-121c-41de-8d3e-59e21dc78cf3", 00:27:35.957 "is_configured": true, 00:27:35.957 "data_offset": 0, 00:27:35.957 "data_size": 65536 00:27:35.957 }, 00:27:35.957 { 00:27:35.957 "name": null, 00:27:35.957 "uuid": "7dcdb139-7abd-4b2e-9f24-4c0e75b17ffc", 00:27:35.957 "is_configured": false, 00:27:35.957 "data_offset": 0, 00:27:35.957 "data_size": 65536 00:27:35.957 }, 00:27:35.957 { 00:27:35.957 "name": null, 00:27:35.957 "uuid": "9b4e0f65-0c94-4839-bc75-a9b125b75b7e", 00:27:35.957 "is_configured": false, 00:27:35.957 "data_offset": 0, 00:27:35.957 "data_size": 65536 00:27:35.957 } 00:27:35.957 ] 00:27:35.957 }' 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:35.957 13:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:36.215 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:36.215 13:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:36.215 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:27:36.215 13:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:36.215 13:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:36.215 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:27:36.215 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:27:36.215 13:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:36.215 13:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:36.215 [2024-11-20 13:47:43.257472] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:27:36.215 13:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:36.215 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:27:36.215 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:36.215 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:36.215 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:36.215 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:36.215 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:36.215 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:36.215 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:36.215 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:36.215 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:36.215 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:36.215 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:36.215 13:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:36.215 13:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:36.472 13:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:36.472 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:36.472 "name": "Existed_Raid", 00:27:36.472 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:36.472 "strip_size_kb": 64, 00:27:36.472 "state": "configuring", 00:27:36.472 "raid_level": "raid0", 00:27:36.472 "superblock": false, 00:27:36.472 "num_base_bdevs": 3, 00:27:36.472 "num_base_bdevs_discovered": 2, 00:27:36.472 "num_base_bdevs_operational": 3, 00:27:36.472 "base_bdevs_list": [ 00:27:36.472 { 00:27:36.472 "name": "BaseBdev1", 00:27:36.472 "uuid": "922d6536-121c-41de-8d3e-59e21dc78cf3", 00:27:36.472 "is_configured": true, 00:27:36.472 "data_offset": 0, 00:27:36.472 "data_size": 65536 00:27:36.472 }, 00:27:36.472 { 00:27:36.472 "name": null, 00:27:36.472 "uuid": "7dcdb139-7abd-4b2e-9f24-4c0e75b17ffc", 00:27:36.472 "is_configured": false, 00:27:36.473 "data_offset": 0, 00:27:36.473 "data_size": 65536 00:27:36.473 }, 00:27:36.473 { 00:27:36.473 "name": "BaseBdev3", 00:27:36.473 "uuid": "9b4e0f65-0c94-4839-bc75-a9b125b75b7e", 00:27:36.473 "is_configured": true, 00:27:36.473 "data_offset": 0, 00:27:36.473 "data_size": 65536 00:27:36.473 } 00:27:36.473 ] 00:27:36.473 }' 00:27:36.473 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:36.473 13:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:36.731 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:36.731 13:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:36.731 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:27:36.731 13:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:36.731 13:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:36.731 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:27:36.731 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:27:36.731 13:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:36.731 13:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:36.731 [2024-11-20 13:47:43.593554] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:27:36.731 13:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:36.731 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:27:36.731 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:36.731 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:36.731 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:36.731 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:36.731 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:36.731 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:36.731 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:36.731 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:36.731 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:36.731 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:36.731 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:36.731 13:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:36.731 13:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:36.731 13:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:36.731 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:36.731 "name": "Existed_Raid", 00:27:36.731 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:36.731 "strip_size_kb": 64, 00:27:36.731 "state": "configuring", 00:27:36.731 "raid_level": "raid0", 00:27:36.731 "superblock": false, 00:27:36.731 "num_base_bdevs": 3, 00:27:36.731 "num_base_bdevs_discovered": 1, 00:27:36.731 "num_base_bdevs_operational": 3, 00:27:36.731 "base_bdevs_list": [ 00:27:36.731 { 00:27:36.731 "name": null, 00:27:36.731 "uuid": "922d6536-121c-41de-8d3e-59e21dc78cf3", 00:27:36.731 "is_configured": false, 00:27:36.731 "data_offset": 0, 00:27:36.731 "data_size": 65536 00:27:36.731 }, 00:27:36.732 { 00:27:36.732 "name": null, 00:27:36.732 "uuid": "7dcdb139-7abd-4b2e-9f24-4c0e75b17ffc", 00:27:36.732 "is_configured": false, 00:27:36.732 "data_offset": 0, 00:27:36.732 "data_size": 65536 00:27:36.732 }, 00:27:36.732 { 00:27:36.732 "name": "BaseBdev3", 00:27:36.732 "uuid": "9b4e0f65-0c94-4839-bc75-a9b125b75b7e", 00:27:36.732 "is_configured": true, 00:27:36.732 "data_offset": 0, 00:27:36.732 "data_size": 65536 00:27:36.732 } 00:27:36.732 ] 00:27:36.732 }' 00:27:36.732 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:36.732 13:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:36.988 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:27:36.988 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:36.988 13:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:36.988 13:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:36.988 13:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:36.988 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:27:36.988 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:27:36.988 13:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:36.988 13:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:36.988 [2024-11-20 13:47:43.995661] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:27:36.988 13:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:36.988 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:27:36.988 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:36.988 13:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:36.988 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:36.988 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:36.988 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:36.988 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:36.988 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:36.988 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:36.988 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:36.988 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:36.988 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:36.988 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:36.988 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:36.988 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:36.988 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:36.988 "name": "Existed_Raid", 00:27:36.988 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:36.988 "strip_size_kb": 64, 00:27:36.988 "state": "configuring", 00:27:36.988 "raid_level": "raid0", 00:27:36.988 "superblock": false, 00:27:36.988 "num_base_bdevs": 3, 00:27:36.988 "num_base_bdevs_discovered": 2, 00:27:36.988 "num_base_bdevs_operational": 3, 00:27:36.988 "base_bdevs_list": [ 00:27:36.988 { 00:27:36.988 "name": null, 00:27:36.989 "uuid": "922d6536-121c-41de-8d3e-59e21dc78cf3", 00:27:36.989 "is_configured": false, 00:27:36.989 "data_offset": 0, 00:27:36.989 "data_size": 65536 00:27:36.989 }, 00:27:36.989 { 00:27:36.989 "name": "BaseBdev2", 00:27:36.989 "uuid": "7dcdb139-7abd-4b2e-9f24-4c0e75b17ffc", 00:27:36.989 "is_configured": true, 00:27:36.989 "data_offset": 0, 00:27:36.989 "data_size": 65536 00:27:36.989 }, 00:27:36.989 { 00:27:36.989 "name": "BaseBdev3", 00:27:36.989 "uuid": "9b4e0f65-0c94-4839-bc75-a9b125b75b7e", 00:27:36.989 "is_configured": true, 00:27:36.989 "data_offset": 0, 00:27:36.989 "data_size": 65536 00:27:36.989 } 00:27:36.989 ] 00:27:36.989 }' 00:27:36.989 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:36.989 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 922d6536-121c-41de-8d3e-59e21dc78cf3 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:37.553 [2024-11-20 13:47:44.428238] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:27:37.553 [2024-11-20 13:47:44.428429] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:27:37.553 [2024-11-20 13:47:44.428445] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:27:37.553 [2024-11-20 13:47:44.428672] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:27:37.553 [2024-11-20 13:47:44.428794] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:27:37.553 [2024-11-20 13:47:44.428801] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:27:37.553 [2024-11-20 13:47:44.429038] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:37.553 NewBaseBdev 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:37.553 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:37.553 [ 00:27:37.553 { 00:27:37.553 "name": "NewBaseBdev", 00:27:37.553 "aliases": [ 00:27:37.553 "922d6536-121c-41de-8d3e-59e21dc78cf3" 00:27:37.553 ], 00:27:37.553 "product_name": "Malloc disk", 00:27:37.553 "block_size": 512, 00:27:37.553 "num_blocks": 65536, 00:27:37.553 "uuid": "922d6536-121c-41de-8d3e-59e21dc78cf3", 00:27:37.554 "assigned_rate_limits": { 00:27:37.554 "rw_ios_per_sec": 0, 00:27:37.554 "rw_mbytes_per_sec": 0, 00:27:37.554 "r_mbytes_per_sec": 0, 00:27:37.554 "w_mbytes_per_sec": 0 00:27:37.554 }, 00:27:37.554 "claimed": true, 00:27:37.554 "claim_type": "exclusive_write", 00:27:37.554 "zoned": false, 00:27:37.554 "supported_io_types": { 00:27:37.554 "read": true, 00:27:37.554 "write": true, 00:27:37.554 "unmap": true, 00:27:37.554 "flush": true, 00:27:37.554 "reset": true, 00:27:37.554 "nvme_admin": false, 00:27:37.554 "nvme_io": false, 00:27:37.554 "nvme_io_md": false, 00:27:37.554 "write_zeroes": true, 00:27:37.554 "zcopy": true, 00:27:37.554 "get_zone_info": false, 00:27:37.554 "zone_management": false, 00:27:37.554 "zone_append": false, 00:27:37.554 "compare": false, 00:27:37.554 "compare_and_write": false, 00:27:37.554 "abort": true, 00:27:37.554 "seek_hole": false, 00:27:37.554 "seek_data": false, 00:27:37.554 "copy": true, 00:27:37.554 "nvme_iov_md": false 00:27:37.554 }, 00:27:37.554 "memory_domains": [ 00:27:37.554 { 00:27:37.554 "dma_device_id": "system", 00:27:37.554 "dma_device_type": 1 00:27:37.554 }, 00:27:37.554 { 00:27:37.554 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:37.554 "dma_device_type": 2 00:27:37.554 } 00:27:37.554 ], 00:27:37.554 "driver_specific": {} 00:27:37.554 } 00:27:37.554 ] 00:27:37.554 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:37.554 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:27:37.554 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:27:37.554 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:37.554 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:37.554 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:37.554 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:37.554 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:37.554 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:37.554 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:37.554 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:37.554 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:37.554 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:37.554 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:37.554 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:37.554 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:37.554 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:37.554 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:37.554 "name": "Existed_Raid", 00:27:37.554 "uuid": "4d459f39-4dd4-481b-897e-8638fcae99b3", 00:27:37.554 "strip_size_kb": 64, 00:27:37.554 "state": "online", 00:27:37.554 "raid_level": "raid0", 00:27:37.554 "superblock": false, 00:27:37.554 "num_base_bdevs": 3, 00:27:37.554 "num_base_bdevs_discovered": 3, 00:27:37.554 "num_base_bdevs_operational": 3, 00:27:37.554 "base_bdevs_list": [ 00:27:37.554 { 00:27:37.554 "name": "NewBaseBdev", 00:27:37.554 "uuid": "922d6536-121c-41de-8d3e-59e21dc78cf3", 00:27:37.554 "is_configured": true, 00:27:37.554 "data_offset": 0, 00:27:37.554 "data_size": 65536 00:27:37.554 }, 00:27:37.554 { 00:27:37.554 "name": "BaseBdev2", 00:27:37.554 "uuid": "7dcdb139-7abd-4b2e-9f24-4c0e75b17ffc", 00:27:37.554 "is_configured": true, 00:27:37.554 "data_offset": 0, 00:27:37.554 "data_size": 65536 00:27:37.554 }, 00:27:37.554 { 00:27:37.554 "name": "BaseBdev3", 00:27:37.554 "uuid": "9b4e0f65-0c94-4839-bc75-a9b125b75b7e", 00:27:37.554 "is_configured": true, 00:27:37.554 "data_offset": 0, 00:27:37.554 "data_size": 65536 00:27:37.554 } 00:27:37.554 ] 00:27:37.554 }' 00:27:37.554 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:37.554 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:37.812 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:27:37.812 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:27:37.812 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:27:37.812 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:27:37.812 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:27:37.812 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:27:37.812 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:27:37.812 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:27:37.812 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:37.812 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:37.812 [2024-11-20 13:47:44.792650] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:37.812 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:37.812 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:27:37.812 "name": "Existed_Raid", 00:27:37.812 "aliases": [ 00:27:37.812 "4d459f39-4dd4-481b-897e-8638fcae99b3" 00:27:37.812 ], 00:27:37.812 "product_name": "Raid Volume", 00:27:37.812 "block_size": 512, 00:27:37.812 "num_blocks": 196608, 00:27:37.812 "uuid": "4d459f39-4dd4-481b-897e-8638fcae99b3", 00:27:37.812 "assigned_rate_limits": { 00:27:37.812 "rw_ios_per_sec": 0, 00:27:37.812 "rw_mbytes_per_sec": 0, 00:27:37.812 "r_mbytes_per_sec": 0, 00:27:37.812 "w_mbytes_per_sec": 0 00:27:37.812 }, 00:27:37.812 "claimed": false, 00:27:37.812 "zoned": false, 00:27:37.812 "supported_io_types": { 00:27:37.812 "read": true, 00:27:37.812 "write": true, 00:27:37.812 "unmap": true, 00:27:37.812 "flush": true, 00:27:37.812 "reset": true, 00:27:37.812 "nvme_admin": false, 00:27:37.812 "nvme_io": false, 00:27:37.812 "nvme_io_md": false, 00:27:37.812 "write_zeroes": true, 00:27:37.812 "zcopy": false, 00:27:37.812 "get_zone_info": false, 00:27:37.812 "zone_management": false, 00:27:37.812 "zone_append": false, 00:27:37.812 "compare": false, 00:27:37.812 "compare_and_write": false, 00:27:37.812 "abort": false, 00:27:37.812 "seek_hole": false, 00:27:37.812 "seek_data": false, 00:27:37.812 "copy": false, 00:27:37.812 "nvme_iov_md": false 00:27:37.812 }, 00:27:37.812 "memory_domains": [ 00:27:37.812 { 00:27:37.812 "dma_device_id": "system", 00:27:37.812 "dma_device_type": 1 00:27:37.812 }, 00:27:37.812 { 00:27:37.812 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:37.812 "dma_device_type": 2 00:27:37.812 }, 00:27:37.812 { 00:27:37.812 "dma_device_id": "system", 00:27:37.812 "dma_device_type": 1 00:27:37.812 }, 00:27:37.812 { 00:27:37.812 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:37.812 "dma_device_type": 2 00:27:37.812 }, 00:27:37.812 { 00:27:37.812 "dma_device_id": "system", 00:27:37.812 "dma_device_type": 1 00:27:37.812 }, 00:27:37.812 { 00:27:37.812 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:37.812 "dma_device_type": 2 00:27:37.812 } 00:27:37.812 ], 00:27:37.812 "driver_specific": { 00:27:37.812 "raid": { 00:27:37.812 "uuid": "4d459f39-4dd4-481b-897e-8638fcae99b3", 00:27:37.812 "strip_size_kb": 64, 00:27:37.812 "state": "online", 00:27:37.812 "raid_level": "raid0", 00:27:37.812 "superblock": false, 00:27:37.812 "num_base_bdevs": 3, 00:27:37.812 "num_base_bdevs_discovered": 3, 00:27:37.812 "num_base_bdevs_operational": 3, 00:27:37.812 "base_bdevs_list": [ 00:27:37.812 { 00:27:37.812 "name": "NewBaseBdev", 00:27:37.812 "uuid": "922d6536-121c-41de-8d3e-59e21dc78cf3", 00:27:37.812 "is_configured": true, 00:27:37.813 "data_offset": 0, 00:27:37.813 "data_size": 65536 00:27:37.813 }, 00:27:37.813 { 00:27:37.813 "name": "BaseBdev2", 00:27:37.813 "uuid": "7dcdb139-7abd-4b2e-9f24-4c0e75b17ffc", 00:27:37.813 "is_configured": true, 00:27:37.813 "data_offset": 0, 00:27:37.813 "data_size": 65536 00:27:37.813 }, 00:27:37.813 { 00:27:37.813 "name": "BaseBdev3", 00:27:37.813 "uuid": "9b4e0f65-0c94-4839-bc75-a9b125b75b7e", 00:27:37.813 "is_configured": true, 00:27:37.813 "data_offset": 0, 00:27:37.813 "data_size": 65536 00:27:37.813 } 00:27:37.813 ] 00:27:37.813 } 00:27:37.813 } 00:27:37.813 }' 00:27:37.813 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:27:37.813 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:27:37.813 BaseBdev2 00:27:37.813 BaseBdev3' 00:27:37.813 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:38.071 [2024-11-20 13:47:44.980414] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:27:38.071 [2024-11-20 13:47:44.980442] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:27:38.071 [2024-11-20 13:47:44.980523] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:38.071 [2024-11-20 13:47:44.980585] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:38.071 [2024-11-20 13:47:44.980596] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 62347 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 62347 ']' 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 62347 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:38.071 13:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 62347 00:27:38.071 killing process with pid 62347 00:27:38.071 13:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:27:38.071 13:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:27:38.071 13:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 62347' 00:27:38.071 13:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 62347 00:27:38.071 [2024-11-20 13:47:45.009745] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:27:38.071 13:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 62347 00:27:38.329 [2024-11-20 13:47:45.168458] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:27:38.894 00:27:38.894 real 0m7.637s 00:27:38.894 user 0m12.116s 00:27:38.894 sys 0m1.345s 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:27:38.894 ************************************ 00:27:38.894 END TEST raid_state_function_test 00:27:38.894 ************************************ 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:38.894 13:47:45 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 3 true 00:27:38.894 13:47:45 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:27:38.894 13:47:45 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:27:38.894 13:47:45 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:27:38.894 ************************************ 00:27:38.894 START TEST raid_state_function_test_sb 00:27:38.894 ************************************ 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 3 true 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:27:38.894 Process raid pid: 62935 00:27:38.894 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:27:38.894 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:27:38.895 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:27:38.895 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=62935 00:27:38.895 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 62935' 00:27:38.895 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 62935 00:27:38.895 13:47:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 62935 ']' 00:27:38.895 13:47:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:38.895 13:47:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:38.895 13:47:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:38.895 13:47:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:38.895 13:47:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:38.895 13:47:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:27:38.895 [2024-11-20 13:47:45.916912] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:27:38.895 [2024-11-20 13:47:45.917065] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:27:39.152 [2024-11-20 13:47:46.070856] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:39.152 [2024-11-20 13:47:46.190007] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:27:39.410 [2024-11-20 13:47:46.342009] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:39.410 [2024-11-20 13:47:46.342057] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:39.747 13:47:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:39.747 13:47:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:27:39.747 13:47:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:27:39.747 13:47:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:39.747 13:47:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:40.016 [2024-11-20 13:47:46.782145] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:27:40.016 [2024-11-20 13:47:46.782207] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:27:40.016 [2024-11-20 13:47:46.782218] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:27:40.016 [2024-11-20 13:47:46.782228] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:27:40.016 [2024-11-20 13:47:46.782235] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:27:40.016 [2024-11-20 13:47:46.782244] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:27:40.016 13:47:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:40.016 13:47:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:27:40.016 13:47:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:40.016 13:47:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:40.016 13:47:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:40.016 13:47:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:40.016 13:47:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:40.016 13:47:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:40.016 13:47:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:40.016 13:47:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:40.016 13:47:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:40.016 13:47:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:40.016 13:47:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:40.016 13:47:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:40.016 13:47:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:40.016 13:47:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:40.016 13:47:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:40.016 "name": "Existed_Raid", 00:27:40.016 "uuid": "fd3ebade-aef7-491d-8ee3-2acf2538996f", 00:27:40.016 "strip_size_kb": 64, 00:27:40.016 "state": "configuring", 00:27:40.016 "raid_level": "raid0", 00:27:40.016 "superblock": true, 00:27:40.016 "num_base_bdevs": 3, 00:27:40.016 "num_base_bdevs_discovered": 0, 00:27:40.016 "num_base_bdevs_operational": 3, 00:27:40.016 "base_bdevs_list": [ 00:27:40.016 { 00:27:40.016 "name": "BaseBdev1", 00:27:40.016 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:40.016 "is_configured": false, 00:27:40.016 "data_offset": 0, 00:27:40.016 "data_size": 0 00:27:40.016 }, 00:27:40.016 { 00:27:40.016 "name": "BaseBdev2", 00:27:40.016 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:40.016 "is_configured": false, 00:27:40.016 "data_offset": 0, 00:27:40.016 "data_size": 0 00:27:40.016 }, 00:27:40.016 { 00:27:40.016 "name": "BaseBdev3", 00:27:40.016 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:40.016 "is_configured": false, 00:27:40.016 "data_offset": 0, 00:27:40.016 "data_size": 0 00:27:40.016 } 00:27:40.016 ] 00:27:40.016 }' 00:27:40.016 13:47:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:40.016 13:47:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:40.275 [2024-11-20 13:47:47.126149] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:27:40.275 [2024-11-20 13:47:47.126189] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:40.275 [2024-11-20 13:47:47.134150] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:27:40.275 [2024-11-20 13:47:47.134317] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:27:40.275 [2024-11-20 13:47:47.134334] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:27:40.275 [2024-11-20 13:47:47.134345] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:27:40.275 [2024-11-20 13:47:47.134351] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:27:40.275 [2024-11-20 13:47:47.134360] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:40.275 [2024-11-20 13:47:47.168859] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:27:40.275 BaseBdev1 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:40.275 [ 00:27:40.275 { 00:27:40.275 "name": "BaseBdev1", 00:27:40.275 "aliases": [ 00:27:40.275 "c4759342-0813-4468-95bc-1da2b12d6036" 00:27:40.275 ], 00:27:40.275 "product_name": "Malloc disk", 00:27:40.275 "block_size": 512, 00:27:40.275 "num_blocks": 65536, 00:27:40.275 "uuid": "c4759342-0813-4468-95bc-1da2b12d6036", 00:27:40.275 "assigned_rate_limits": { 00:27:40.275 "rw_ios_per_sec": 0, 00:27:40.275 "rw_mbytes_per_sec": 0, 00:27:40.275 "r_mbytes_per_sec": 0, 00:27:40.275 "w_mbytes_per_sec": 0 00:27:40.275 }, 00:27:40.275 "claimed": true, 00:27:40.275 "claim_type": "exclusive_write", 00:27:40.275 "zoned": false, 00:27:40.275 "supported_io_types": { 00:27:40.275 "read": true, 00:27:40.275 "write": true, 00:27:40.275 "unmap": true, 00:27:40.275 "flush": true, 00:27:40.275 "reset": true, 00:27:40.275 "nvme_admin": false, 00:27:40.275 "nvme_io": false, 00:27:40.275 "nvme_io_md": false, 00:27:40.275 "write_zeroes": true, 00:27:40.275 "zcopy": true, 00:27:40.275 "get_zone_info": false, 00:27:40.275 "zone_management": false, 00:27:40.275 "zone_append": false, 00:27:40.275 "compare": false, 00:27:40.275 "compare_and_write": false, 00:27:40.275 "abort": true, 00:27:40.275 "seek_hole": false, 00:27:40.275 "seek_data": false, 00:27:40.275 "copy": true, 00:27:40.275 "nvme_iov_md": false 00:27:40.275 }, 00:27:40.275 "memory_domains": [ 00:27:40.275 { 00:27:40.275 "dma_device_id": "system", 00:27:40.275 "dma_device_type": 1 00:27:40.275 }, 00:27:40.275 { 00:27:40.275 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:40.275 "dma_device_type": 2 00:27:40.275 } 00:27:40.275 ], 00:27:40.275 "driver_specific": {} 00:27:40.275 } 00:27:40.275 ] 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:40.275 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:40.276 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:40.276 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:40.276 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:40.276 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:40.276 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:40.276 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:40.276 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:40.276 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:40.276 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:40.276 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:40.276 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:40.276 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:40.276 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:40.276 "name": "Existed_Raid", 00:27:40.276 "uuid": "c2476a1d-a748-4d96-b29b-06c400c2cf98", 00:27:40.276 "strip_size_kb": 64, 00:27:40.276 "state": "configuring", 00:27:40.276 "raid_level": "raid0", 00:27:40.276 "superblock": true, 00:27:40.276 "num_base_bdevs": 3, 00:27:40.276 "num_base_bdevs_discovered": 1, 00:27:40.276 "num_base_bdevs_operational": 3, 00:27:40.276 "base_bdevs_list": [ 00:27:40.276 { 00:27:40.276 "name": "BaseBdev1", 00:27:40.276 "uuid": "c4759342-0813-4468-95bc-1da2b12d6036", 00:27:40.276 "is_configured": true, 00:27:40.276 "data_offset": 2048, 00:27:40.276 "data_size": 63488 00:27:40.276 }, 00:27:40.276 { 00:27:40.276 "name": "BaseBdev2", 00:27:40.276 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:40.276 "is_configured": false, 00:27:40.276 "data_offset": 0, 00:27:40.276 "data_size": 0 00:27:40.276 }, 00:27:40.276 { 00:27:40.276 "name": "BaseBdev3", 00:27:40.276 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:40.276 "is_configured": false, 00:27:40.276 "data_offset": 0, 00:27:40.276 "data_size": 0 00:27:40.276 } 00:27:40.276 ] 00:27:40.276 }' 00:27:40.276 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:40.276 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:40.534 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:27:40.534 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:40.534 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:40.534 [2024-11-20 13:47:47.497009] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:27:40.534 [2024-11-20 13:47:47.497197] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:27:40.534 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:40.534 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:27:40.534 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:40.534 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:40.534 [2024-11-20 13:47:47.505067] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:27:40.534 [2024-11-20 13:47:47.507077] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:27:40.534 [2024-11-20 13:47:47.507115] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:27:40.534 [2024-11-20 13:47:47.507125] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:27:40.534 [2024-11-20 13:47:47.507134] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:27:40.534 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:40.534 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:27:40.534 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:27:40.534 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:27:40.534 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:40.534 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:40.534 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:40.534 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:40.534 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:40.534 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:40.534 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:40.534 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:40.534 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:40.534 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:40.534 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:40.534 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:40.534 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:40.534 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:40.534 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:40.534 "name": "Existed_Raid", 00:27:40.534 "uuid": "388f8451-1604-4da7-9a75-b6881b9d6311", 00:27:40.534 "strip_size_kb": 64, 00:27:40.534 "state": "configuring", 00:27:40.534 "raid_level": "raid0", 00:27:40.534 "superblock": true, 00:27:40.534 "num_base_bdevs": 3, 00:27:40.534 "num_base_bdevs_discovered": 1, 00:27:40.534 "num_base_bdevs_operational": 3, 00:27:40.534 "base_bdevs_list": [ 00:27:40.534 { 00:27:40.534 "name": "BaseBdev1", 00:27:40.534 "uuid": "c4759342-0813-4468-95bc-1da2b12d6036", 00:27:40.534 "is_configured": true, 00:27:40.534 "data_offset": 2048, 00:27:40.534 "data_size": 63488 00:27:40.534 }, 00:27:40.534 { 00:27:40.535 "name": "BaseBdev2", 00:27:40.535 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:40.535 "is_configured": false, 00:27:40.535 "data_offset": 0, 00:27:40.535 "data_size": 0 00:27:40.535 }, 00:27:40.535 { 00:27:40.535 "name": "BaseBdev3", 00:27:40.535 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:40.535 "is_configured": false, 00:27:40.535 "data_offset": 0, 00:27:40.535 "data_size": 0 00:27:40.535 } 00:27:40.535 ] 00:27:40.535 }' 00:27:40.535 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:40.535 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:40.793 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:27:40.793 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:40.793 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:40.793 [2024-11-20 13:47:47.825639] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:27:40.793 BaseBdev2 00:27:40.793 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:40.793 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:27:40.793 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:27:40.793 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:27:40.793 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:27:40.793 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:27:40.793 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:27:40.793 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:27:40.793 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:40.793 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:40.793 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:40.793 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:27:40.793 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:40.793 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:41.051 [ 00:27:41.051 { 00:27:41.051 "name": "BaseBdev2", 00:27:41.051 "aliases": [ 00:27:41.051 "f9c13549-c148-4fee-9969-739a9aacd808" 00:27:41.051 ], 00:27:41.051 "product_name": "Malloc disk", 00:27:41.051 "block_size": 512, 00:27:41.051 "num_blocks": 65536, 00:27:41.051 "uuid": "f9c13549-c148-4fee-9969-739a9aacd808", 00:27:41.051 "assigned_rate_limits": { 00:27:41.051 "rw_ios_per_sec": 0, 00:27:41.051 "rw_mbytes_per_sec": 0, 00:27:41.051 "r_mbytes_per_sec": 0, 00:27:41.051 "w_mbytes_per_sec": 0 00:27:41.051 }, 00:27:41.051 "claimed": true, 00:27:41.051 "claim_type": "exclusive_write", 00:27:41.051 "zoned": false, 00:27:41.051 "supported_io_types": { 00:27:41.051 "read": true, 00:27:41.051 "write": true, 00:27:41.051 "unmap": true, 00:27:41.051 "flush": true, 00:27:41.051 "reset": true, 00:27:41.051 "nvme_admin": false, 00:27:41.051 "nvme_io": false, 00:27:41.051 "nvme_io_md": false, 00:27:41.051 "write_zeroes": true, 00:27:41.051 "zcopy": true, 00:27:41.051 "get_zone_info": false, 00:27:41.051 "zone_management": false, 00:27:41.051 "zone_append": false, 00:27:41.051 "compare": false, 00:27:41.051 "compare_and_write": false, 00:27:41.051 "abort": true, 00:27:41.051 "seek_hole": false, 00:27:41.051 "seek_data": false, 00:27:41.051 "copy": true, 00:27:41.051 "nvme_iov_md": false 00:27:41.051 }, 00:27:41.051 "memory_domains": [ 00:27:41.051 { 00:27:41.051 "dma_device_id": "system", 00:27:41.051 "dma_device_type": 1 00:27:41.051 }, 00:27:41.051 { 00:27:41.051 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:41.051 "dma_device_type": 2 00:27:41.051 } 00:27:41.051 ], 00:27:41.051 "driver_specific": {} 00:27:41.051 } 00:27:41.051 ] 00:27:41.051 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:41.051 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:27:41.051 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:27:41.051 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:27:41.051 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:27:41.051 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:41.051 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:41.051 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:41.051 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:41.051 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:41.051 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:41.051 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:41.051 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:41.051 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:41.051 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:41.051 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:41.051 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:41.051 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:41.051 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:41.051 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:41.051 "name": "Existed_Raid", 00:27:41.051 "uuid": "388f8451-1604-4da7-9a75-b6881b9d6311", 00:27:41.051 "strip_size_kb": 64, 00:27:41.051 "state": "configuring", 00:27:41.051 "raid_level": "raid0", 00:27:41.051 "superblock": true, 00:27:41.051 "num_base_bdevs": 3, 00:27:41.051 "num_base_bdevs_discovered": 2, 00:27:41.051 "num_base_bdevs_operational": 3, 00:27:41.051 "base_bdevs_list": [ 00:27:41.051 { 00:27:41.051 "name": "BaseBdev1", 00:27:41.051 "uuid": "c4759342-0813-4468-95bc-1da2b12d6036", 00:27:41.051 "is_configured": true, 00:27:41.051 "data_offset": 2048, 00:27:41.051 "data_size": 63488 00:27:41.051 }, 00:27:41.051 { 00:27:41.051 "name": "BaseBdev2", 00:27:41.051 "uuid": "f9c13549-c148-4fee-9969-739a9aacd808", 00:27:41.051 "is_configured": true, 00:27:41.051 "data_offset": 2048, 00:27:41.051 "data_size": 63488 00:27:41.051 }, 00:27:41.051 { 00:27:41.051 "name": "BaseBdev3", 00:27:41.052 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:41.052 "is_configured": false, 00:27:41.052 "data_offset": 0, 00:27:41.052 "data_size": 0 00:27:41.052 } 00:27:41.052 ] 00:27:41.052 }' 00:27:41.052 13:47:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:41.052 13:47:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:41.310 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:27:41.310 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:41.310 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:41.310 [2024-11-20 13:47:48.185242] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:27:41.310 [2024-11-20 13:47:48.185724] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:27:41.310 [2024-11-20 13:47:48.185833] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:27:41.310 BaseBdev3 00:27:41.310 [2024-11-20 13:47:48.186152] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:27:41.310 [2024-11-20 13:47:48.186302] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:27:41.310 [2024-11-20 13:47:48.186311] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:27:41.310 [2024-11-20 13:47:48.186452] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:41.310 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:41.310 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:27:41.310 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:27:41.310 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:27:41.310 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:27:41.310 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:27:41.310 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:27:41.310 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:27:41.310 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:41.310 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:41.310 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:41.310 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:27:41.310 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:41.310 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:41.310 [ 00:27:41.310 { 00:27:41.310 "name": "BaseBdev3", 00:27:41.310 "aliases": [ 00:27:41.310 "3f85f115-f612-4535-95d6-65c81b60b821" 00:27:41.310 ], 00:27:41.310 "product_name": "Malloc disk", 00:27:41.310 "block_size": 512, 00:27:41.310 "num_blocks": 65536, 00:27:41.310 "uuid": "3f85f115-f612-4535-95d6-65c81b60b821", 00:27:41.310 "assigned_rate_limits": { 00:27:41.310 "rw_ios_per_sec": 0, 00:27:41.310 "rw_mbytes_per_sec": 0, 00:27:41.310 "r_mbytes_per_sec": 0, 00:27:41.310 "w_mbytes_per_sec": 0 00:27:41.310 }, 00:27:41.310 "claimed": true, 00:27:41.310 "claim_type": "exclusive_write", 00:27:41.310 "zoned": false, 00:27:41.310 "supported_io_types": { 00:27:41.311 "read": true, 00:27:41.311 "write": true, 00:27:41.311 "unmap": true, 00:27:41.311 "flush": true, 00:27:41.311 "reset": true, 00:27:41.311 "nvme_admin": false, 00:27:41.311 "nvme_io": false, 00:27:41.311 "nvme_io_md": false, 00:27:41.311 "write_zeroes": true, 00:27:41.311 "zcopy": true, 00:27:41.311 "get_zone_info": false, 00:27:41.311 "zone_management": false, 00:27:41.311 "zone_append": false, 00:27:41.311 "compare": false, 00:27:41.311 "compare_and_write": false, 00:27:41.311 "abort": true, 00:27:41.311 "seek_hole": false, 00:27:41.311 "seek_data": false, 00:27:41.311 "copy": true, 00:27:41.311 "nvme_iov_md": false 00:27:41.311 }, 00:27:41.311 "memory_domains": [ 00:27:41.311 { 00:27:41.311 "dma_device_id": "system", 00:27:41.311 "dma_device_type": 1 00:27:41.311 }, 00:27:41.311 { 00:27:41.311 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:41.311 "dma_device_type": 2 00:27:41.311 } 00:27:41.311 ], 00:27:41.311 "driver_specific": {} 00:27:41.311 } 00:27:41.311 ] 00:27:41.311 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:41.311 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:27:41.311 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:27:41.311 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:27:41.311 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:27:41.311 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:41.311 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:41.311 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:41.311 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:41.311 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:41.311 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:41.311 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:41.311 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:41.311 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:41.311 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:41.311 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:41.311 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:41.311 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:41.311 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:41.311 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:41.311 "name": "Existed_Raid", 00:27:41.311 "uuid": "388f8451-1604-4da7-9a75-b6881b9d6311", 00:27:41.311 "strip_size_kb": 64, 00:27:41.311 "state": "online", 00:27:41.311 "raid_level": "raid0", 00:27:41.311 "superblock": true, 00:27:41.311 "num_base_bdevs": 3, 00:27:41.311 "num_base_bdevs_discovered": 3, 00:27:41.311 "num_base_bdevs_operational": 3, 00:27:41.311 "base_bdevs_list": [ 00:27:41.311 { 00:27:41.311 "name": "BaseBdev1", 00:27:41.311 "uuid": "c4759342-0813-4468-95bc-1da2b12d6036", 00:27:41.311 "is_configured": true, 00:27:41.311 "data_offset": 2048, 00:27:41.311 "data_size": 63488 00:27:41.311 }, 00:27:41.311 { 00:27:41.311 "name": "BaseBdev2", 00:27:41.311 "uuid": "f9c13549-c148-4fee-9969-739a9aacd808", 00:27:41.311 "is_configured": true, 00:27:41.311 "data_offset": 2048, 00:27:41.311 "data_size": 63488 00:27:41.311 }, 00:27:41.311 { 00:27:41.311 "name": "BaseBdev3", 00:27:41.311 "uuid": "3f85f115-f612-4535-95d6-65c81b60b821", 00:27:41.311 "is_configured": true, 00:27:41.311 "data_offset": 2048, 00:27:41.311 "data_size": 63488 00:27:41.311 } 00:27:41.311 ] 00:27:41.311 }' 00:27:41.311 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:41.311 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:41.570 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:27:41.570 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:27:41.570 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:27:41.570 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:27:41.570 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:27:41.570 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:27:41.570 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:27:41.570 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:41.570 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:41.570 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:27:41.570 [2024-11-20 13:47:48.529723] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:41.570 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:41.570 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:27:41.570 "name": "Existed_Raid", 00:27:41.570 "aliases": [ 00:27:41.570 "388f8451-1604-4da7-9a75-b6881b9d6311" 00:27:41.570 ], 00:27:41.570 "product_name": "Raid Volume", 00:27:41.570 "block_size": 512, 00:27:41.570 "num_blocks": 190464, 00:27:41.570 "uuid": "388f8451-1604-4da7-9a75-b6881b9d6311", 00:27:41.570 "assigned_rate_limits": { 00:27:41.570 "rw_ios_per_sec": 0, 00:27:41.570 "rw_mbytes_per_sec": 0, 00:27:41.570 "r_mbytes_per_sec": 0, 00:27:41.570 "w_mbytes_per_sec": 0 00:27:41.570 }, 00:27:41.570 "claimed": false, 00:27:41.570 "zoned": false, 00:27:41.570 "supported_io_types": { 00:27:41.570 "read": true, 00:27:41.570 "write": true, 00:27:41.570 "unmap": true, 00:27:41.570 "flush": true, 00:27:41.570 "reset": true, 00:27:41.570 "nvme_admin": false, 00:27:41.570 "nvme_io": false, 00:27:41.570 "nvme_io_md": false, 00:27:41.570 "write_zeroes": true, 00:27:41.570 "zcopy": false, 00:27:41.570 "get_zone_info": false, 00:27:41.570 "zone_management": false, 00:27:41.570 "zone_append": false, 00:27:41.570 "compare": false, 00:27:41.570 "compare_and_write": false, 00:27:41.570 "abort": false, 00:27:41.570 "seek_hole": false, 00:27:41.570 "seek_data": false, 00:27:41.570 "copy": false, 00:27:41.570 "nvme_iov_md": false 00:27:41.570 }, 00:27:41.570 "memory_domains": [ 00:27:41.570 { 00:27:41.570 "dma_device_id": "system", 00:27:41.570 "dma_device_type": 1 00:27:41.570 }, 00:27:41.570 { 00:27:41.570 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:41.570 "dma_device_type": 2 00:27:41.570 }, 00:27:41.570 { 00:27:41.570 "dma_device_id": "system", 00:27:41.570 "dma_device_type": 1 00:27:41.570 }, 00:27:41.570 { 00:27:41.570 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:41.570 "dma_device_type": 2 00:27:41.570 }, 00:27:41.570 { 00:27:41.570 "dma_device_id": "system", 00:27:41.570 "dma_device_type": 1 00:27:41.570 }, 00:27:41.570 { 00:27:41.570 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:41.570 "dma_device_type": 2 00:27:41.570 } 00:27:41.570 ], 00:27:41.570 "driver_specific": { 00:27:41.570 "raid": { 00:27:41.570 "uuid": "388f8451-1604-4da7-9a75-b6881b9d6311", 00:27:41.570 "strip_size_kb": 64, 00:27:41.570 "state": "online", 00:27:41.570 "raid_level": "raid0", 00:27:41.570 "superblock": true, 00:27:41.570 "num_base_bdevs": 3, 00:27:41.570 "num_base_bdevs_discovered": 3, 00:27:41.570 "num_base_bdevs_operational": 3, 00:27:41.570 "base_bdevs_list": [ 00:27:41.570 { 00:27:41.570 "name": "BaseBdev1", 00:27:41.570 "uuid": "c4759342-0813-4468-95bc-1da2b12d6036", 00:27:41.570 "is_configured": true, 00:27:41.570 "data_offset": 2048, 00:27:41.570 "data_size": 63488 00:27:41.570 }, 00:27:41.570 { 00:27:41.570 "name": "BaseBdev2", 00:27:41.570 "uuid": "f9c13549-c148-4fee-9969-739a9aacd808", 00:27:41.570 "is_configured": true, 00:27:41.570 "data_offset": 2048, 00:27:41.570 "data_size": 63488 00:27:41.570 }, 00:27:41.570 { 00:27:41.570 "name": "BaseBdev3", 00:27:41.570 "uuid": "3f85f115-f612-4535-95d6-65c81b60b821", 00:27:41.570 "is_configured": true, 00:27:41.570 "data_offset": 2048, 00:27:41.570 "data_size": 63488 00:27:41.570 } 00:27:41.570 ] 00:27:41.570 } 00:27:41.570 } 00:27:41.570 }' 00:27:41.570 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:27:41.570 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:27:41.570 BaseBdev2 00:27:41.570 BaseBdev3' 00:27:41.570 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:41.828 [2024-11-20 13:47:48.733495] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:27:41.828 [2024-11-20 13:47:48.733527] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:27:41.828 [2024-11-20 13:47:48.733589] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 2 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:41.828 "name": "Existed_Raid", 00:27:41.828 "uuid": "388f8451-1604-4da7-9a75-b6881b9d6311", 00:27:41.828 "strip_size_kb": 64, 00:27:41.828 "state": "offline", 00:27:41.828 "raid_level": "raid0", 00:27:41.828 "superblock": true, 00:27:41.828 "num_base_bdevs": 3, 00:27:41.828 "num_base_bdevs_discovered": 2, 00:27:41.828 "num_base_bdevs_operational": 2, 00:27:41.828 "base_bdevs_list": [ 00:27:41.828 { 00:27:41.828 "name": null, 00:27:41.828 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:41.828 "is_configured": false, 00:27:41.828 "data_offset": 0, 00:27:41.828 "data_size": 63488 00:27:41.828 }, 00:27:41.828 { 00:27:41.828 "name": "BaseBdev2", 00:27:41.828 "uuid": "f9c13549-c148-4fee-9969-739a9aacd808", 00:27:41.828 "is_configured": true, 00:27:41.828 "data_offset": 2048, 00:27:41.828 "data_size": 63488 00:27:41.828 }, 00:27:41.828 { 00:27:41.828 "name": "BaseBdev3", 00:27:41.828 "uuid": "3f85f115-f612-4535-95d6-65c81b60b821", 00:27:41.828 "is_configured": true, 00:27:41.828 "data_offset": 2048, 00:27:41.828 "data_size": 63488 00:27:41.828 } 00:27:41.828 ] 00:27:41.828 }' 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:41.828 13:47:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:42.086 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:27:42.086 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:27:42.086 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:42.086 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:42.086 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:42.086 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:27:42.086 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:42.086 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:27:42.086 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:27:42.086 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:27:42.086 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:42.086 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:42.086 [2024-11-20 13:47:49.139126] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:42.342 [2024-11-20 13:47:49.253919] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:27:42.342 [2024-11-20 13:47:49.253991] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:42.342 BaseBdev2 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:42.342 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:42.600 [ 00:27:42.600 { 00:27:42.600 "name": "BaseBdev2", 00:27:42.600 "aliases": [ 00:27:42.600 "e13a2c5b-6338-4934-a1f3-b5252081270a" 00:27:42.600 ], 00:27:42.600 "product_name": "Malloc disk", 00:27:42.600 "block_size": 512, 00:27:42.600 "num_blocks": 65536, 00:27:42.600 "uuid": "e13a2c5b-6338-4934-a1f3-b5252081270a", 00:27:42.600 "assigned_rate_limits": { 00:27:42.600 "rw_ios_per_sec": 0, 00:27:42.600 "rw_mbytes_per_sec": 0, 00:27:42.600 "r_mbytes_per_sec": 0, 00:27:42.600 "w_mbytes_per_sec": 0 00:27:42.600 }, 00:27:42.600 "claimed": false, 00:27:42.600 "zoned": false, 00:27:42.600 "supported_io_types": { 00:27:42.600 "read": true, 00:27:42.600 "write": true, 00:27:42.600 "unmap": true, 00:27:42.600 "flush": true, 00:27:42.600 "reset": true, 00:27:42.600 "nvme_admin": false, 00:27:42.600 "nvme_io": false, 00:27:42.600 "nvme_io_md": false, 00:27:42.600 "write_zeroes": true, 00:27:42.600 "zcopy": true, 00:27:42.600 "get_zone_info": false, 00:27:42.600 "zone_management": false, 00:27:42.600 "zone_append": false, 00:27:42.600 "compare": false, 00:27:42.600 "compare_and_write": false, 00:27:42.600 "abort": true, 00:27:42.600 "seek_hole": false, 00:27:42.600 "seek_data": false, 00:27:42.600 "copy": true, 00:27:42.600 "nvme_iov_md": false 00:27:42.600 }, 00:27:42.600 "memory_domains": [ 00:27:42.600 { 00:27:42.600 "dma_device_id": "system", 00:27:42.600 "dma_device_type": 1 00:27:42.600 }, 00:27:42.600 { 00:27:42.600 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:42.600 "dma_device_type": 2 00:27:42.600 } 00:27:42.600 ], 00:27:42.600 "driver_specific": {} 00:27:42.600 } 00:27:42.600 ] 00:27:42.600 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:42.600 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:27:42.600 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:27:42.600 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:27:42.600 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:27:42.600 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:42.600 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:42.600 BaseBdev3 00:27:42.600 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:42.600 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:27:42.600 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:27:42.600 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:27:42.600 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:27:42.600 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:27:42.600 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:27:42.600 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:27:42.600 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:42.600 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:42.600 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:42.600 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:27:42.600 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:42.600 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:42.600 [ 00:27:42.600 { 00:27:42.600 "name": "BaseBdev3", 00:27:42.600 "aliases": [ 00:27:42.600 "39196627-e645-4b07-a23b-c5074a8cf5ce" 00:27:42.600 ], 00:27:42.600 "product_name": "Malloc disk", 00:27:42.600 "block_size": 512, 00:27:42.600 "num_blocks": 65536, 00:27:42.600 "uuid": "39196627-e645-4b07-a23b-c5074a8cf5ce", 00:27:42.600 "assigned_rate_limits": { 00:27:42.600 "rw_ios_per_sec": 0, 00:27:42.600 "rw_mbytes_per_sec": 0, 00:27:42.600 "r_mbytes_per_sec": 0, 00:27:42.600 "w_mbytes_per_sec": 0 00:27:42.600 }, 00:27:42.600 "claimed": false, 00:27:42.600 "zoned": false, 00:27:42.600 "supported_io_types": { 00:27:42.600 "read": true, 00:27:42.600 "write": true, 00:27:42.600 "unmap": true, 00:27:42.600 "flush": true, 00:27:42.600 "reset": true, 00:27:42.601 "nvme_admin": false, 00:27:42.601 "nvme_io": false, 00:27:42.601 "nvme_io_md": false, 00:27:42.601 "write_zeroes": true, 00:27:42.601 "zcopy": true, 00:27:42.601 "get_zone_info": false, 00:27:42.601 "zone_management": false, 00:27:42.601 "zone_append": false, 00:27:42.601 "compare": false, 00:27:42.601 "compare_and_write": false, 00:27:42.601 "abort": true, 00:27:42.601 "seek_hole": false, 00:27:42.601 "seek_data": false, 00:27:42.601 "copy": true, 00:27:42.601 "nvme_iov_md": false 00:27:42.601 }, 00:27:42.601 "memory_domains": [ 00:27:42.601 { 00:27:42.601 "dma_device_id": "system", 00:27:42.601 "dma_device_type": 1 00:27:42.601 }, 00:27:42.601 { 00:27:42.601 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:42.601 "dma_device_type": 2 00:27:42.601 } 00:27:42.601 ], 00:27:42.601 "driver_specific": {} 00:27:42.601 } 00:27:42.601 ] 00:27:42.601 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:42.601 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:27:42.601 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:27:42.601 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:27:42.601 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:27:42.601 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:42.601 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:42.601 [2024-11-20 13:47:49.474046] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:27:42.601 [2024-11-20 13:47:49.474212] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:27:42.601 [2024-11-20 13:47:49.474289] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:27:42.601 [2024-11-20 13:47:49.476380] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:27:42.601 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:42.601 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:27:42.601 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:42.601 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:42.601 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:42.601 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:42.601 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:42.601 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:42.601 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:42.601 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:42.601 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:42.601 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:42.601 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:42.601 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:42.601 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:42.601 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:42.601 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:42.601 "name": "Existed_Raid", 00:27:42.601 "uuid": "4d2d2b26-a65d-4473-8e17-e63283a6010d", 00:27:42.601 "strip_size_kb": 64, 00:27:42.601 "state": "configuring", 00:27:42.601 "raid_level": "raid0", 00:27:42.601 "superblock": true, 00:27:42.601 "num_base_bdevs": 3, 00:27:42.601 "num_base_bdevs_discovered": 2, 00:27:42.601 "num_base_bdevs_operational": 3, 00:27:42.601 "base_bdevs_list": [ 00:27:42.601 { 00:27:42.601 "name": "BaseBdev1", 00:27:42.601 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:42.601 "is_configured": false, 00:27:42.601 "data_offset": 0, 00:27:42.601 "data_size": 0 00:27:42.601 }, 00:27:42.601 { 00:27:42.601 "name": "BaseBdev2", 00:27:42.601 "uuid": "e13a2c5b-6338-4934-a1f3-b5252081270a", 00:27:42.601 "is_configured": true, 00:27:42.601 "data_offset": 2048, 00:27:42.601 "data_size": 63488 00:27:42.601 }, 00:27:42.601 { 00:27:42.601 "name": "BaseBdev3", 00:27:42.601 "uuid": "39196627-e645-4b07-a23b-c5074a8cf5ce", 00:27:42.601 "is_configured": true, 00:27:42.601 "data_offset": 2048, 00:27:42.601 "data_size": 63488 00:27:42.601 } 00:27:42.601 ] 00:27:42.601 }' 00:27:42.601 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:42.601 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:42.858 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:27:42.858 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:42.858 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:42.858 [2024-11-20 13:47:49.806122] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:27:42.858 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:42.858 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:27:42.858 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:42.858 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:42.858 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:42.858 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:42.858 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:42.858 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:42.858 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:42.858 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:42.858 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:42.858 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:42.858 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:42.858 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:42.858 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:42.858 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:42.858 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:42.858 "name": "Existed_Raid", 00:27:42.858 "uuid": "4d2d2b26-a65d-4473-8e17-e63283a6010d", 00:27:42.858 "strip_size_kb": 64, 00:27:42.858 "state": "configuring", 00:27:42.858 "raid_level": "raid0", 00:27:42.858 "superblock": true, 00:27:42.858 "num_base_bdevs": 3, 00:27:42.858 "num_base_bdevs_discovered": 1, 00:27:42.858 "num_base_bdevs_operational": 3, 00:27:42.858 "base_bdevs_list": [ 00:27:42.858 { 00:27:42.858 "name": "BaseBdev1", 00:27:42.858 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:42.859 "is_configured": false, 00:27:42.859 "data_offset": 0, 00:27:42.859 "data_size": 0 00:27:42.859 }, 00:27:42.859 { 00:27:42.859 "name": null, 00:27:42.859 "uuid": "e13a2c5b-6338-4934-a1f3-b5252081270a", 00:27:42.859 "is_configured": false, 00:27:42.859 "data_offset": 0, 00:27:42.859 "data_size": 63488 00:27:42.859 }, 00:27:42.859 { 00:27:42.859 "name": "BaseBdev3", 00:27:42.859 "uuid": "39196627-e645-4b07-a23b-c5074a8cf5ce", 00:27:42.859 "is_configured": true, 00:27:42.859 "data_offset": 2048, 00:27:42.859 "data_size": 63488 00:27:42.859 } 00:27:42.859 ] 00:27:42.859 }' 00:27:42.859 13:47:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:42.859 13:47:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:43.116 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:43.116 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:27:43.116 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:43.116 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:43.116 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:43.374 [2024-11-20 13:47:50.210973] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:27:43.374 BaseBdev1 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:43.374 [ 00:27:43.374 { 00:27:43.374 "name": "BaseBdev1", 00:27:43.374 "aliases": [ 00:27:43.374 "401ac9b6-f06f-451e-bd04-2366b7a14599" 00:27:43.374 ], 00:27:43.374 "product_name": "Malloc disk", 00:27:43.374 "block_size": 512, 00:27:43.374 "num_blocks": 65536, 00:27:43.374 "uuid": "401ac9b6-f06f-451e-bd04-2366b7a14599", 00:27:43.374 "assigned_rate_limits": { 00:27:43.374 "rw_ios_per_sec": 0, 00:27:43.374 "rw_mbytes_per_sec": 0, 00:27:43.374 "r_mbytes_per_sec": 0, 00:27:43.374 "w_mbytes_per_sec": 0 00:27:43.374 }, 00:27:43.374 "claimed": true, 00:27:43.374 "claim_type": "exclusive_write", 00:27:43.374 "zoned": false, 00:27:43.374 "supported_io_types": { 00:27:43.374 "read": true, 00:27:43.374 "write": true, 00:27:43.374 "unmap": true, 00:27:43.374 "flush": true, 00:27:43.374 "reset": true, 00:27:43.374 "nvme_admin": false, 00:27:43.374 "nvme_io": false, 00:27:43.374 "nvme_io_md": false, 00:27:43.374 "write_zeroes": true, 00:27:43.374 "zcopy": true, 00:27:43.374 "get_zone_info": false, 00:27:43.374 "zone_management": false, 00:27:43.374 "zone_append": false, 00:27:43.374 "compare": false, 00:27:43.374 "compare_and_write": false, 00:27:43.374 "abort": true, 00:27:43.374 "seek_hole": false, 00:27:43.374 "seek_data": false, 00:27:43.374 "copy": true, 00:27:43.374 "nvme_iov_md": false 00:27:43.374 }, 00:27:43.374 "memory_domains": [ 00:27:43.374 { 00:27:43.374 "dma_device_id": "system", 00:27:43.374 "dma_device_type": 1 00:27:43.374 }, 00:27:43.374 { 00:27:43.374 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:43.374 "dma_device_type": 2 00:27:43.374 } 00:27:43.374 ], 00:27:43.374 "driver_specific": {} 00:27:43.374 } 00:27:43.374 ] 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:43.374 "name": "Existed_Raid", 00:27:43.374 "uuid": "4d2d2b26-a65d-4473-8e17-e63283a6010d", 00:27:43.374 "strip_size_kb": 64, 00:27:43.374 "state": "configuring", 00:27:43.374 "raid_level": "raid0", 00:27:43.374 "superblock": true, 00:27:43.374 "num_base_bdevs": 3, 00:27:43.374 "num_base_bdevs_discovered": 2, 00:27:43.374 "num_base_bdevs_operational": 3, 00:27:43.374 "base_bdevs_list": [ 00:27:43.374 { 00:27:43.374 "name": "BaseBdev1", 00:27:43.374 "uuid": "401ac9b6-f06f-451e-bd04-2366b7a14599", 00:27:43.374 "is_configured": true, 00:27:43.374 "data_offset": 2048, 00:27:43.374 "data_size": 63488 00:27:43.374 }, 00:27:43.374 { 00:27:43.374 "name": null, 00:27:43.374 "uuid": "e13a2c5b-6338-4934-a1f3-b5252081270a", 00:27:43.374 "is_configured": false, 00:27:43.374 "data_offset": 0, 00:27:43.374 "data_size": 63488 00:27:43.374 }, 00:27:43.374 { 00:27:43.374 "name": "BaseBdev3", 00:27:43.374 "uuid": "39196627-e645-4b07-a23b-c5074a8cf5ce", 00:27:43.374 "is_configured": true, 00:27:43.374 "data_offset": 2048, 00:27:43.374 "data_size": 63488 00:27:43.374 } 00:27:43.374 ] 00:27:43.374 }' 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:43.374 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:43.664 [2024-11-20 13:47:50.607103] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:43.664 "name": "Existed_Raid", 00:27:43.664 "uuid": "4d2d2b26-a65d-4473-8e17-e63283a6010d", 00:27:43.664 "strip_size_kb": 64, 00:27:43.664 "state": "configuring", 00:27:43.664 "raid_level": "raid0", 00:27:43.664 "superblock": true, 00:27:43.664 "num_base_bdevs": 3, 00:27:43.664 "num_base_bdevs_discovered": 1, 00:27:43.664 "num_base_bdevs_operational": 3, 00:27:43.664 "base_bdevs_list": [ 00:27:43.664 { 00:27:43.664 "name": "BaseBdev1", 00:27:43.664 "uuid": "401ac9b6-f06f-451e-bd04-2366b7a14599", 00:27:43.664 "is_configured": true, 00:27:43.664 "data_offset": 2048, 00:27:43.664 "data_size": 63488 00:27:43.664 }, 00:27:43.664 { 00:27:43.664 "name": null, 00:27:43.664 "uuid": "e13a2c5b-6338-4934-a1f3-b5252081270a", 00:27:43.664 "is_configured": false, 00:27:43.664 "data_offset": 0, 00:27:43.664 "data_size": 63488 00:27:43.664 }, 00:27:43.664 { 00:27:43.664 "name": null, 00:27:43.664 "uuid": "39196627-e645-4b07-a23b-c5074a8cf5ce", 00:27:43.664 "is_configured": false, 00:27:43.664 "data_offset": 0, 00:27:43.664 "data_size": 63488 00:27:43.664 } 00:27:43.664 ] 00:27:43.664 }' 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:43.664 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:43.947 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:43.947 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:27:43.947 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:43.947 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:43.947 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:43.947 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:27:43.947 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:27:43.947 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:43.947 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:43.947 [2024-11-20 13:47:50.955224] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:27:43.947 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:43.947 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:27:43.947 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:43.947 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:43.947 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:43.947 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:43.947 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:43.947 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:43.947 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:43.947 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:43.947 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:43.947 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:43.947 13:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:43.947 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:43.947 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:43.947 13:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:44.206 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:44.206 "name": "Existed_Raid", 00:27:44.206 "uuid": "4d2d2b26-a65d-4473-8e17-e63283a6010d", 00:27:44.206 "strip_size_kb": 64, 00:27:44.206 "state": "configuring", 00:27:44.207 "raid_level": "raid0", 00:27:44.207 "superblock": true, 00:27:44.207 "num_base_bdevs": 3, 00:27:44.207 "num_base_bdevs_discovered": 2, 00:27:44.207 "num_base_bdevs_operational": 3, 00:27:44.207 "base_bdevs_list": [ 00:27:44.207 { 00:27:44.207 "name": "BaseBdev1", 00:27:44.207 "uuid": "401ac9b6-f06f-451e-bd04-2366b7a14599", 00:27:44.207 "is_configured": true, 00:27:44.207 "data_offset": 2048, 00:27:44.207 "data_size": 63488 00:27:44.207 }, 00:27:44.207 { 00:27:44.207 "name": null, 00:27:44.207 "uuid": "e13a2c5b-6338-4934-a1f3-b5252081270a", 00:27:44.207 "is_configured": false, 00:27:44.207 "data_offset": 0, 00:27:44.207 "data_size": 63488 00:27:44.207 }, 00:27:44.207 { 00:27:44.207 "name": "BaseBdev3", 00:27:44.207 "uuid": "39196627-e645-4b07-a23b-c5074a8cf5ce", 00:27:44.207 "is_configured": true, 00:27:44.207 "data_offset": 2048, 00:27:44.207 "data_size": 63488 00:27:44.207 } 00:27:44.207 ] 00:27:44.207 }' 00:27:44.207 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:44.207 13:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:44.466 [2024-11-20 13:47:51.315307] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:44.466 "name": "Existed_Raid", 00:27:44.466 "uuid": "4d2d2b26-a65d-4473-8e17-e63283a6010d", 00:27:44.466 "strip_size_kb": 64, 00:27:44.466 "state": "configuring", 00:27:44.466 "raid_level": "raid0", 00:27:44.466 "superblock": true, 00:27:44.466 "num_base_bdevs": 3, 00:27:44.466 "num_base_bdevs_discovered": 1, 00:27:44.466 "num_base_bdevs_operational": 3, 00:27:44.466 "base_bdevs_list": [ 00:27:44.466 { 00:27:44.466 "name": null, 00:27:44.466 "uuid": "401ac9b6-f06f-451e-bd04-2366b7a14599", 00:27:44.466 "is_configured": false, 00:27:44.466 "data_offset": 0, 00:27:44.466 "data_size": 63488 00:27:44.466 }, 00:27:44.466 { 00:27:44.466 "name": null, 00:27:44.466 "uuid": "e13a2c5b-6338-4934-a1f3-b5252081270a", 00:27:44.466 "is_configured": false, 00:27:44.466 "data_offset": 0, 00:27:44.466 "data_size": 63488 00:27:44.466 }, 00:27:44.466 { 00:27:44.466 "name": "BaseBdev3", 00:27:44.466 "uuid": "39196627-e645-4b07-a23b-c5074a8cf5ce", 00:27:44.466 "is_configured": true, 00:27:44.466 "data_offset": 2048, 00:27:44.466 "data_size": 63488 00:27:44.466 } 00:27:44.466 ] 00:27:44.466 }' 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:44.466 13:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:44.725 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:44.725 13:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:44.725 13:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:44.725 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:27:44.725 13:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:44.725 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:27:44.725 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:27:44.725 13:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:44.725 13:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:44.725 [2024-11-20 13:47:51.749829] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:27:44.725 13:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:44.725 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:27:44.725 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:44.725 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:44.725 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:44.725 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:44.725 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:44.725 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:44.725 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:44.725 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:44.725 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:44.725 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:44.725 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:44.725 13:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:44.725 13:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:44.725 13:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:44.984 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:44.984 "name": "Existed_Raid", 00:27:44.984 "uuid": "4d2d2b26-a65d-4473-8e17-e63283a6010d", 00:27:44.984 "strip_size_kb": 64, 00:27:44.984 "state": "configuring", 00:27:44.984 "raid_level": "raid0", 00:27:44.984 "superblock": true, 00:27:44.984 "num_base_bdevs": 3, 00:27:44.984 "num_base_bdevs_discovered": 2, 00:27:44.984 "num_base_bdevs_operational": 3, 00:27:44.984 "base_bdevs_list": [ 00:27:44.984 { 00:27:44.984 "name": null, 00:27:44.984 "uuid": "401ac9b6-f06f-451e-bd04-2366b7a14599", 00:27:44.984 "is_configured": false, 00:27:44.984 "data_offset": 0, 00:27:44.984 "data_size": 63488 00:27:44.984 }, 00:27:44.984 { 00:27:44.984 "name": "BaseBdev2", 00:27:44.984 "uuid": "e13a2c5b-6338-4934-a1f3-b5252081270a", 00:27:44.984 "is_configured": true, 00:27:44.984 "data_offset": 2048, 00:27:44.984 "data_size": 63488 00:27:44.984 }, 00:27:44.984 { 00:27:44.984 "name": "BaseBdev3", 00:27:44.984 "uuid": "39196627-e645-4b07-a23b-c5074a8cf5ce", 00:27:44.984 "is_configured": true, 00:27:44.984 "data_offset": 2048, 00:27:44.984 "data_size": 63488 00:27:44.984 } 00:27:44.984 ] 00:27:44.984 }' 00:27:44.984 13:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:44.984 13:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:45.242 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:45.242 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:27:45.242 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:45.242 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:45.242 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:45.242 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 401ac9b6-f06f-451e-bd04-2366b7a14599 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:45.243 [2024-11-20 13:47:52.174845] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:27:45.243 NewBaseBdev 00:27:45.243 [2024-11-20 13:47:52.175369] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:27:45.243 [2024-11-20 13:47:52.175397] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:27:45.243 [2024-11-20 13:47:52.175685] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:27:45.243 [2024-11-20 13:47:52.175827] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:27:45.243 [2024-11-20 13:47:52.175837] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:27:45.243 [2024-11-20 13:47:52.176010] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:45.243 [ 00:27:45.243 { 00:27:45.243 "name": "NewBaseBdev", 00:27:45.243 "aliases": [ 00:27:45.243 "401ac9b6-f06f-451e-bd04-2366b7a14599" 00:27:45.243 ], 00:27:45.243 "product_name": "Malloc disk", 00:27:45.243 "block_size": 512, 00:27:45.243 "num_blocks": 65536, 00:27:45.243 "uuid": "401ac9b6-f06f-451e-bd04-2366b7a14599", 00:27:45.243 "assigned_rate_limits": { 00:27:45.243 "rw_ios_per_sec": 0, 00:27:45.243 "rw_mbytes_per_sec": 0, 00:27:45.243 "r_mbytes_per_sec": 0, 00:27:45.243 "w_mbytes_per_sec": 0 00:27:45.243 }, 00:27:45.243 "claimed": true, 00:27:45.243 "claim_type": "exclusive_write", 00:27:45.243 "zoned": false, 00:27:45.243 "supported_io_types": { 00:27:45.243 "read": true, 00:27:45.243 "write": true, 00:27:45.243 "unmap": true, 00:27:45.243 "flush": true, 00:27:45.243 "reset": true, 00:27:45.243 "nvme_admin": false, 00:27:45.243 "nvme_io": false, 00:27:45.243 "nvme_io_md": false, 00:27:45.243 "write_zeroes": true, 00:27:45.243 "zcopy": true, 00:27:45.243 "get_zone_info": false, 00:27:45.243 "zone_management": false, 00:27:45.243 "zone_append": false, 00:27:45.243 "compare": false, 00:27:45.243 "compare_and_write": false, 00:27:45.243 "abort": true, 00:27:45.243 "seek_hole": false, 00:27:45.243 "seek_data": false, 00:27:45.243 "copy": true, 00:27:45.243 "nvme_iov_md": false 00:27:45.243 }, 00:27:45.243 "memory_domains": [ 00:27:45.243 { 00:27:45.243 "dma_device_id": "system", 00:27:45.243 "dma_device_type": 1 00:27:45.243 }, 00:27:45.243 { 00:27:45.243 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:45.243 "dma_device_type": 2 00:27:45.243 } 00:27:45.243 ], 00:27:45.243 "driver_specific": {} 00:27:45.243 } 00:27:45.243 ] 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:45.243 "name": "Existed_Raid", 00:27:45.243 "uuid": "4d2d2b26-a65d-4473-8e17-e63283a6010d", 00:27:45.243 "strip_size_kb": 64, 00:27:45.243 "state": "online", 00:27:45.243 "raid_level": "raid0", 00:27:45.243 "superblock": true, 00:27:45.243 "num_base_bdevs": 3, 00:27:45.243 "num_base_bdevs_discovered": 3, 00:27:45.243 "num_base_bdevs_operational": 3, 00:27:45.243 "base_bdevs_list": [ 00:27:45.243 { 00:27:45.243 "name": "NewBaseBdev", 00:27:45.243 "uuid": "401ac9b6-f06f-451e-bd04-2366b7a14599", 00:27:45.243 "is_configured": true, 00:27:45.243 "data_offset": 2048, 00:27:45.243 "data_size": 63488 00:27:45.243 }, 00:27:45.243 { 00:27:45.243 "name": "BaseBdev2", 00:27:45.243 "uuid": "e13a2c5b-6338-4934-a1f3-b5252081270a", 00:27:45.243 "is_configured": true, 00:27:45.243 "data_offset": 2048, 00:27:45.243 "data_size": 63488 00:27:45.243 }, 00:27:45.243 { 00:27:45.243 "name": "BaseBdev3", 00:27:45.243 "uuid": "39196627-e645-4b07-a23b-c5074a8cf5ce", 00:27:45.243 "is_configured": true, 00:27:45.243 "data_offset": 2048, 00:27:45.243 "data_size": 63488 00:27:45.243 } 00:27:45.243 ] 00:27:45.243 }' 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:45.243 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:45.501 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:27:45.501 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:27:45.501 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:27:45.501 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:27:45.501 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:27:45.501 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:27:45.501 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:27:45.501 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:27:45.501 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:45.501 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:45.501 [2024-11-20 13:47:52.515336] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:45.501 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:45.501 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:27:45.501 "name": "Existed_Raid", 00:27:45.501 "aliases": [ 00:27:45.501 "4d2d2b26-a65d-4473-8e17-e63283a6010d" 00:27:45.501 ], 00:27:45.501 "product_name": "Raid Volume", 00:27:45.501 "block_size": 512, 00:27:45.501 "num_blocks": 190464, 00:27:45.501 "uuid": "4d2d2b26-a65d-4473-8e17-e63283a6010d", 00:27:45.501 "assigned_rate_limits": { 00:27:45.501 "rw_ios_per_sec": 0, 00:27:45.501 "rw_mbytes_per_sec": 0, 00:27:45.501 "r_mbytes_per_sec": 0, 00:27:45.501 "w_mbytes_per_sec": 0 00:27:45.501 }, 00:27:45.501 "claimed": false, 00:27:45.501 "zoned": false, 00:27:45.501 "supported_io_types": { 00:27:45.501 "read": true, 00:27:45.501 "write": true, 00:27:45.501 "unmap": true, 00:27:45.501 "flush": true, 00:27:45.501 "reset": true, 00:27:45.501 "nvme_admin": false, 00:27:45.501 "nvme_io": false, 00:27:45.501 "nvme_io_md": false, 00:27:45.501 "write_zeroes": true, 00:27:45.501 "zcopy": false, 00:27:45.501 "get_zone_info": false, 00:27:45.501 "zone_management": false, 00:27:45.501 "zone_append": false, 00:27:45.501 "compare": false, 00:27:45.501 "compare_and_write": false, 00:27:45.501 "abort": false, 00:27:45.501 "seek_hole": false, 00:27:45.501 "seek_data": false, 00:27:45.501 "copy": false, 00:27:45.501 "nvme_iov_md": false 00:27:45.501 }, 00:27:45.501 "memory_domains": [ 00:27:45.501 { 00:27:45.501 "dma_device_id": "system", 00:27:45.501 "dma_device_type": 1 00:27:45.501 }, 00:27:45.501 { 00:27:45.501 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:45.501 "dma_device_type": 2 00:27:45.501 }, 00:27:45.501 { 00:27:45.501 "dma_device_id": "system", 00:27:45.501 "dma_device_type": 1 00:27:45.501 }, 00:27:45.501 { 00:27:45.501 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:45.501 "dma_device_type": 2 00:27:45.501 }, 00:27:45.501 { 00:27:45.501 "dma_device_id": "system", 00:27:45.501 "dma_device_type": 1 00:27:45.501 }, 00:27:45.501 { 00:27:45.501 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:45.501 "dma_device_type": 2 00:27:45.501 } 00:27:45.501 ], 00:27:45.501 "driver_specific": { 00:27:45.501 "raid": { 00:27:45.501 "uuid": "4d2d2b26-a65d-4473-8e17-e63283a6010d", 00:27:45.501 "strip_size_kb": 64, 00:27:45.501 "state": "online", 00:27:45.501 "raid_level": "raid0", 00:27:45.501 "superblock": true, 00:27:45.501 "num_base_bdevs": 3, 00:27:45.501 "num_base_bdevs_discovered": 3, 00:27:45.501 "num_base_bdevs_operational": 3, 00:27:45.501 "base_bdevs_list": [ 00:27:45.501 { 00:27:45.501 "name": "NewBaseBdev", 00:27:45.502 "uuid": "401ac9b6-f06f-451e-bd04-2366b7a14599", 00:27:45.502 "is_configured": true, 00:27:45.502 "data_offset": 2048, 00:27:45.502 "data_size": 63488 00:27:45.502 }, 00:27:45.502 { 00:27:45.502 "name": "BaseBdev2", 00:27:45.502 "uuid": "e13a2c5b-6338-4934-a1f3-b5252081270a", 00:27:45.502 "is_configured": true, 00:27:45.502 "data_offset": 2048, 00:27:45.502 "data_size": 63488 00:27:45.502 }, 00:27:45.502 { 00:27:45.502 "name": "BaseBdev3", 00:27:45.502 "uuid": "39196627-e645-4b07-a23b-c5074a8cf5ce", 00:27:45.502 "is_configured": true, 00:27:45.502 "data_offset": 2048, 00:27:45.502 "data_size": 63488 00:27:45.502 } 00:27:45.502 ] 00:27:45.502 } 00:27:45.502 } 00:27:45.502 }' 00:27:45.502 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:27:45.760 BaseBdev2 00:27:45.760 BaseBdev3' 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:45.760 [2024-11-20 13:47:52.711045] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:27:45.760 [2024-11-20 13:47:52.711084] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:27:45.760 [2024-11-20 13:47:52.711170] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:45.760 [2024-11-20 13:47:52.711242] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:45.760 [2024-11-20 13:47:52.711255] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 62935 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 62935 ']' 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 62935 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 62935 00:27:45.760 killing process with pid 62935 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 62935' 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 62935 00:27:45.760 [2024-11-20 13:47:52.742341] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:27:45.760 13:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 62935 00:27:46.018 [2024-11-20 13:47:52.941860] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:27:46.959 13:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:27:46.959 00:27:46.959 real 0m7.857s 00:27:46.959 user 0m12.411s 00:27:46.959 sys 0m1.349s 00:27:46.959 ************************************ 00:27:46.959 END TEST raid_state_function_test_sb 00:27:46.959 ************************************ 00:27:46.959 13:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:27:46.959 13:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:46.959 13:47:53 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 3 00:27:46.959 13:47:53 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:27:46.959 13:47:53 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:27:46.959 13:47:53 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:27:46.959 ************************************ 00:27:46.959 START TEST raid_superblock_test 00:27:46.959 ************************************ 00:27:46.959 13:47:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid0 3 00:27:46.959 13:47:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:27:46.959 13:47:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:27:46.959 13:47:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:27:46.959 13:47:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:27:46.959 13:47:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:27:46.959 13:47:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:27:46.959 13:47:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:27:46.959 13:47:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:27:46.959 13:47:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:27:46.959 13:47:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:27:46.959 13:47:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:27:46.959 13:47:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:27:46.959 13:47:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:27:46.959 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:46.959 13:47:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:27:46.959 13:47:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:27:46.959 13:47:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:27:46.959 13:47:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=63533 00:27:46.959 13:47:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 63533 00:27:46.959 13:47:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 63533 ']' 00:27:46.959 13:47:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:46.959 13:47:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:46.959 13:47:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:46.959 13:47:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:46.959 13:47:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:46.959 13:47:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:27:46.959 [2024-11-20 13:47:53.814560] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:27:46.959 [2024-11-20 13:47:53.814845] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid63533 ] 00:27:46.959 [2024-11-20 13:47:53.975311] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:47.251 [2024-11-20 13:47:54.094487] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:27:47.251 [2024-11-20 13:47:54.244216] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:47.251 [2024-11-20 13:47:54.244458] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:47.824 malloc1 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:47.824 [2024-11-20 13:47:54.638486] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:27:47.824 [2024-11-20 13:47:54.638558] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:47.824 [2024-11-20 13:47:54.638582] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:27:47.824 [2024-11-20 13:47:54.638592] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:47.824 [2024-11-20 13:47:54.640943] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:47.824 [2024-11-20 13:47:54.640996] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:27:47.824 pt1 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:47.824 malloc2 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:47.824 [2024-11-20 13:47:54.676857] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:27:47.824 [2024-11-20 13:47:54.676928] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:47.824 [2024-11-20 13:47:54.676958] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:27:47.824 [2024-11-20 13:47:54.676968] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:47.824 [2024-11-20 13:47:54.679300] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:47.824 [2024-11-20 13:47:54.679338] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:27:47.824 pt2 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:47.824 malloc3 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:47.824 [2024-11-20 13:47:54.730899] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:27:47.824 [2024-11-20 13:47:54.730970] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:47.824 [2024-11-20 13:47:54.731013] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:27:47.824 [2024-11-20 13:47:54.731024] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:47.824 [2024-11-20 13:47:54.733325] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:47.824 [2024-11-20 13:47:54.733362] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:27:47.824 pt3 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:47.824 [2024-11-20 13:47:54.738964] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:27:47.824 [2024-11-20 13:47:54.741080] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:27:47.824 [2024-11-20 13:47:54.741156] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:27:47.824 [2024-11-20 13:47:54.741330] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:27:47.824 [2024-11-20 13:47:54.741343] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:27:47.824 [2024-11-20 13:47:54.741650] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:27:47.824 [2024-11-20 13:47:54.741813] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:27:47.824 [2024-11-20 13:47:54.741830] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:27:47.824 [2024-11-20 13:47:54.742020] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:47.824 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:47.824 "name": "raid_bdev1", 00:27:47.824 "uuid": "bbb8aa17-176c-49cb-ade7-e8e1ff17ac37", 00:27:47.824 "strip_size_kb": 64, 00:27:47.824 "state": "online", 00:27:47.824 "raid_level": "raid0", 00:27:47.824 "superblock": true, 00:27:47.824 "num_base_bdevs": 3, 00:27:47.824 "num_base_bdevs_discovered": 3, 00:27:47.825 "num_base_bdevs_operational": 3, 00:27:47.825 "base_bdevs_list": [ 00:27:47.825 { 00:27:47.825 "name": "pt1", 00:27:47.825 "uuid": "00000000-0000-0000-0000-000000000001", 00:27:47.825 "is_configured": true, 00:27:47.825 "data_offset": 2048, 00:27:47.825 "data_size": 63488 00:27:47.825 }, 00:27:47.825 { 00:27:47.825 "name": "pt2", 00:27:47.825 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:47.825 "is_configured": true, 00:27:47.825 "data_offset": 2048, 00:27:47.825 "data_size": 63488 00:27:47.825 }, 00:27:47.825 { 00:27:47.825 "name": "pt3", 00:27:47.825 "uuid": "00000000-0000-0000-0000-000000000003", 00:27:47.825 "is_configured": true, 00:27:47.825 "data_offset": 2048, 00:27:47.825 "data_size": 63488 00:27:47.825 } 00:27:47.825 ] 00:27:47.825 }' 00:27:47.825 13:47:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:47.825 13:47:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:48.083 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:27:48.083 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:27:48.083 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:27:48.083 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:27:48.083 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:27:48.083 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:27:48.083 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:27:48.083 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:27:48.083 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:48.083 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:48.083 [2024-11-20 13:47:55.079365] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:48.083 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:48.083 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:27:48.083 "name": "raid_bdev1", 00:27:48.083 "aliases": [ 00:27:48.083 "bbb8aa17-176c-49cb-ade7-e8e1ff17ac37" 00:27:48.083 ], 00:27:48.083 "product_name": "Raid Volume", 00:27:48.083 "block_size": 512, 00:27:48.083 "num_blocks": 190464, 00:27:48.083 "uuid": "bbb8aa17-176c-49cb-ade7-e8e1ff17ac37", 00:27:48.083 "assigned_rate_limits": { 00:27:48.083 "rw_ios_per_sec": 0, 00:27:48.083 "rw_mbytes_per_sec": 0, 00:27:48.083 "r_mbytes_per_sec": 0, 00:27:48.083 "w_mbytes_per_sec": 0 00:27:48.083 }, 00:27:48.083 "claimed": false, 00:27:48.083 "zoned": false, 00:27:48.083 "supported_io_types": { 00:27:48.083 "read": true, 00:27:48.083 "write": true, 00:27:48.083 "unmap": true, 00:27:48.083 "flush": true, 00:27:48.083 "reset": true, 00:27:48.083 "nvme_admin": false, 00:27:48.083 "nvme_io": false, 00:27:48.083 "nvme_io_md": false, 00:27:48.083 "write_zeroes": true, 00:27:48.083 "zcopy": false, 00:27:48.083 "get_zone_info": false, 00:27:48.083 "zone_management": false, 00:27:48.083 "zone_append": false, 00:27:48.083 "compare": false, 00:27:48.083 "compare_and_write": false, 00:27:48.083 "abort": false, 00:27:48.083 "seek_hole": false, 00:27:48.083 "seek_data": false, 00:27:48.083 "copy": false, 00:27:48.083 "nvme_iov_md": false 00:27:48.083 }, 00:27:48.083 "memory_domains": [ 00:27:48.083 { 00:27:48.083 "dma_device_id": "system", 00:27:48.083 "dma_device_type": 1 00:27:48.083 }, 00:27:48.083 { 00:27:48.083 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:48.083 "dma_device_type": 2 00:27:48.083 }, 00:27:48.083 { 00:27:48.083 "dma_device_id": "system", 00:27:48.083 "dma_device_type": 1 00:27:48.083 }, 00:27:48.083 { 00:27:48.083 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:48.083 "dma_device_type": 2 00:27:48.083 }, 00:27:48.083 { 00:27:48.083 "dma_device_id": "system", 00:27:48.083 "dma_device_type": 1 00:27:48.083 }, 00:27:48.083 { 00:27:48.083 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:48.083 "dma_device_type": 2 00:27:48.083 } 00:27:48.083 ], 00:27:48.083 "driver_specific": { 00:27:48.083 "raid": { 00:27:48.083 "uuid": "bbb8aa17-176c-49cb-ade7-e8e1ff17ac37", 00:27:48.083 "strip_size_kb": 64, 00:27:48.083 "state": "online", 00:27:48.083 "raid_level": "raid0", 00:27:48.083 "superblock": true, 00:27:48.083 "num_base_bdevs": 3, 00:27:48.083 "num_base_bdevs_discovered": 3, 00:27:48.083 "num_base_bdevs_operational": 3, 00:27:48.083 "base_bdevs_list": [ 00:27:48.083 { 00:27:48.083 "name": "pt1", 00:27:48.083 "uuid": "00000000-0000-0000-0000-000000000001", 00:27:48.083 "is_configured": true, 00:27:48.083 "data_offset": 2048, 00:27:48.083 "data_size": 63488 00:27:48.083 }, 00:27:48.083 { 00:27:48.083 "name": "pt2", 00:27:48.083 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:48.083 "is_configured": true, 00:27:48.083 "data_offset": 2048, 00:27:48.083 "data_size": 63488 00:27:48.083 }, 00:27:48.083 { 00:27:48.083 "name": "pt3", 00:27:48.083 "uuid": "00000000-0000-0000-0000-000000000003", 00:27:48.083 "is_configured": true, 00:27:48.083 "data_offset": 2048, 00:27:48.083 "data_size": 63488 00:27:48.083 } 00:27:48.083 ] 00:27:48.083 } 00:27:48.083 } 00:27:48.083 }' 00:27:48.083 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:27:48.083 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:27:48.083 pt2 00:27:48.083 pt3' 00:27:48.083 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:48.341 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:27:48.341 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:48.341 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:27:48.341 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:48.341 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:48.341 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:48.341 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:48.341 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:48.341 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:48.341 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:48.341 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:48.341 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:27:48.341 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:48.341 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:48.341 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:48.341 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:48.341 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:48.341 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:48.341 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:48.342 [2024-11-20 13:47:55.267355] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=bbb8aa17-176c-49cb-ade7-e8e1ff17ac37 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z bbb8aa17-176c-49cb-ade7-e8e1ff17ac37 ']' 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:48.342 [2024-11-20 13:47:55.295053] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:27:48.342 [2024-11-20 13:47:55.295092] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:27:48.342 [2024-11-20 13:47:55.295182] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:48.342 [2024-11-20 13:47:55.295259] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:48.342 [2024-11-20 13:47:55.295270] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:48.342 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:48.342 [2024-11-20 13:47:55.395132] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:27:48.342 [2024-11-20 13:47:55.397377] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:27:48.342 [2024-11-20 13:47:55.397504] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:27:48.601 [2024-11-20 13:47:55.397629] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:27:48.601 [2024-11-20 13:47:55.397769] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:27:48.601 [2024-11-20 13:47:55.397855] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:27:48.601 [2024-11-20 13:47:55.398048] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:27:48.601 [2024-11-20 13:47:55.398077] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:27:48.601 request: 00:27:48.601 { 00:27:48.601 "name": "raid_bdev1", 00:27:48.601 "raid_level": "raid0", 00:27:48.601 "base_bdevs": [ 00:27:48.601 "malloc1", 00:27:48.601 "malloc2", 00:27:48.601 "malloc3" 00:27:48.601 ], 00:27:48.601 "strip_size_kb": 64, 00:27:48.601 "superblock": false, 00:27:48.601 "method": "bdev_raid_create", 00:27:48.601 "req_id": 1 00:27:48.601 } 00:27:48.601 Got JSON-RPC error response 00:27:48.601 response: 00:27:48.601 { 00:27:48.601 "code": -17, 00:27:48.601 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:27:48.601 } 00:27:48.601 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:27:48.601 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:27:48.601 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:27:48.601 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:27:48.601 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:27:48.601 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:48.601 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:27:48.601 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:48.601 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:48.601 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:48.601 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:27:48.601 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:27:48.601 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:27:48.601 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:48.601 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:48.601 [2024-11-20 13:47:55.443103] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:27:48.601 [2024-11-20 13:47:55.443183] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:48.601 [2024-11-20 13:47:55.443205] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:27:48.601 [2024-11-20 13:47:55.443215] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:48.601 [2024-11-20 13:47:55.445638] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:48.601 [2024-11-20 13:47:55.445681] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:27:48.601 [2024-11-20 13:47:55.445772] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:27:48.601 [2024-11-20 13:47:55.445824] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:27:48.601 pt1 00:27:48.601 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:48.601 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 3 00:27:48.602 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:48.602 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:48.602 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:48.602 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:48.602 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:48.602 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:48.602 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:48.602 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:48.602 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:48.602 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:48.602 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:48.602 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:48.602 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:48.602 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:48.602 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:48.602 "name": "raid_bdev1", 00:27:48.602 "uuid": "bbb8aa17-176c-49cb-ade7-e8e1ff17ac37", 00:27:48.602 "strip_size_kb": 64, 00:27:48.602 "state": "configuring", 00:27:48.602 "raid_level": "raid0", 00:27:48.602 "superblock": true, 00:27:48.602 "num_base_bdevs": 3, 00:27:48.602 "num_base_bdevs_discovered": 1, 00:27:48.602 "num_base_bdevs_operational": 3, 00:27:48.602 "base_bdevs_list": [ 00:27:48.602 { 00:27:48.602 "name": "pt1", 00:27:48.602 "uuid": "00000000-0000-0000-0000-000000000001", 00:27:48.602 "is_configured": true, 00:27:48.602 "data_offset": 2048, 00:27:48.602 "data_size": 63488 00:27:48.602 }, 00:27:48.602 { 00:27:48.602 "name": null, 00:27:48.602 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:48.602 "is_configured": false, 00:27:48.602 "data_offset": 2048, 00:27:48.602 "data_size": 63488 00:27:48.602 }, 00:27:48.602 { 00:27:48.602 "name": null, 00:27:48.602 "uuid": "00000000-0000-0000-0000-000000000003", 00:27:48.602 "is_configured": false, 00:27:48.602 "data_offset": 2048, 00:27:48.602 "data_size": 63488 00:27:48.602 } 00:27:48.602 ] 00:27:48.602 }' 00:27:48.602 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:48.602 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:48.860 [2024-11-20 13:47:55.731165] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:27:48.860 [2024-11-20 13:47:55.731247] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:48.860 [2024-11-20 13:47:55.731271] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:27:48.860 [2024-11-20 13:47:55.731281] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:48.860 [2024-11-20 13:47:55.731750] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:48.860 [2024-11-20 13:47:55.731779] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:27:48.860 [2024-11-20 13:47:55.731870] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:27:48.860 [2024-11-20 13:47:55.731896] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:27:48.860 pt2 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:48.860 [2024-11-20 13:47:55.739189] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 3 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:48.860 "name": "raid_bdev1", 00:27:48.860 "uuid": "bbb8aa17-176c-49cb-ade7-e8e1ff17ac37", 00:27:48.860 "strip_size_kb": 64, 00:27:48.860 "state": "configuring", 00:27:48.860 "raid_level": "raid0", 00:27:48.860 "superblock": true, 00:27:48.860 "num_base_bdevs": 3, 00:27:48.860 "num_base_bdevs_discovered": 1, 00:27:48.860 "num_base_bdevs_operational": 3, 00:27:48.860 "base_bdevs_list": [ 00:27:48.860 { 00:27:48.860 "name": "pt1", 00:27:48.860 "uuid": "00000000-0000-0000-0000-000000000001", 00:27:48.860 "is_configured": true, 00:27:48.860 "data_offset": 2048, 00:27:48.860 "data_size": 63488 00:27:48.860 }, 00:27:48.860 { 00:27:48.860 "name": null, 00:27:48.860 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:48.860 "is_configured": false, 00:27:48.860 "data_offset": 0, 00:27:48.860 "data_size": 63488 00:27:48.860 }, 00:27:48.860 { 00:27:48.860 "name": null, 00:27:48.860 "uuid": "00000000-0000-0000-0000-000000000003", 00:27:48.860 "is_configured": false, 00:27:48.860 "data_offset": 2048, 00:27:48.860 "data_size": 63488 00:27:48.860 } 00:27:48.860 ] 00:27:48.860 }' 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:48.860 13:47:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:49.118 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:27:49.118 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:27:49.118 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:27:49.118 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:49.118 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:49.118 [2024-11-20 13:47:56.051221] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:27:49.118 [2024-11-20 13:47:56.051304] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:49.118 [2024-11-20 13:47:56.051323] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:27:49.118 [2024-11-20 13:47:56.051334] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:49.118 [2024-11-20 13:47:56.051824] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:49.118 [2024-11-20 13:47:56.051848] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:27:49.118 [2024-11-20 13:47:56.051930] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:27:49.118 [2024-11-20 13:47:56.051953] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:27:49.118 pt2 00:27:49.118 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:49.118 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:27:49.118 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:27:49.118 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:27:49.118 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:49.118 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:49.118 [2024-11-20 13:47:56.059225] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:27:49.118 [2024-11-20 13:47:56.059287] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:49.118 [2024-11-20 13:47:56.059304] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:27:49.118 [2024-11-20 13:47:56.059316] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:49.118 [2024-11-20 13:47:56.059773] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:49.119 [2024-11-20 13:47:56.059803] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:27:49.119 [2024-11-20 13:47:56.059879] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:27:49.119 [2024-11-20 13:47:56.059901] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:27:49.119 [2024-11-20 13:47:56.060044] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:27:49.119 [2024-11-20 13:47:56.060252] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:27:49.119 [2024-11-20 13:47:56.060528] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:27:49.119 [2024-11-20 13:47:56.060675] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:27:49.119 [2024-11-20 13:47:56.060684] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:27:49.119 [2024-11-20 13:47:56.060826] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:49.119 pt3 00:27:49.119 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:49.119 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:27:49.119 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:27:49.119 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:27:49.119 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:49.119 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:49.119 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:49.119 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:49.119 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:49.119 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:49.119 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:49.119 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:49.119 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:49.119 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:49.119 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:49.119 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:49.119 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:49.119 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:49.119 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:49.119 "name": "raid_bdev1", 00:27:49.119 "uuid": "bbb8aa17-176c-49cb-ade7-e8e1ff17ac37", 00:27:49.119 "strip_size_kb": 64, 00:27:49.119 "state": "online", 00:27:49.119 "raid_level": "raid0", 00:27:49.119 "superblock": true, 00:27:49.119 "num_base_bdevs": 3, 00:27:49.119 "num_base_bdevs_discovered": 3, 00:27:49.119 "num_base_bdevs_operational": 3, 00:27:49.119 "base_bdevs_list": [ 00:27:49.119 { 00:27:49.119 "name": "pt1", 00:27:49.119 "uuid": "00000000-0000-0000-0000-000000000001", 00:27:49.119 "is_configured": true, 00:27:49.119 "data_offset": 2048, 00:27:49.119 "data_size": 63488 00:27:49.119 }, 00:27:49.119 { 00:27:49.119 "name": "pt2", 00:27:49.119 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:49.119 "is_configured": true, 00:27:49.119 "data_offset": 2048, 00:27:49.119 "data_size": 63488 00:27:49.119 }, 00:27:49.119 { 00:27:49.119 "name": "pt3", 00:27:49.119 "uuid": "00000000-0000-0000-0000-000000000003", 00:27:49.119 "is_configured": true, 00:27:49.119 "data_offset": 2048, 00:27:49.119 "data_size": 63488 00:27:49.119 } 00:27:49.119 ] 00:27:49.119 }' 00:27:49.119 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:49.119 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:49.377 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:27:49.377 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:27:49.377 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:27:49.377 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:27:49.377 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:27:49.377 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:27:49.377 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:27:49.377 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:27:49.377 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:49.377 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:49.377 [2024-11-20 13:47:56.391693] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:49.377 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:49.377 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:27:49.377 "name": "raid_bdev1", 00:27:49.377 "aliases": [ 00:27:49.377 "bbb8aa17-176c-49cb-ade7-e8e1ff17ac37" 00:27:49.377 ], 00:27:49.377 "product_name": "Raid Volume", 00:27:49.377 "block_size": 512, 00:27:49.377 "num_blocks": 190464, 00:27:49.377 "uuid": "bbb8aa17-176c-49cb-ade7-e8e1ff17ac37", 00:27:49.377 "assigned_rate_limits": { 00:27:49.378 "rw_ios_per_sec": 0, 00:27:49.378 "rw_mbytes_per_sec": 0, 00:27:49.378 "r_mbytes_per_sec": 0, 00:27:49.378 "w_mbytes_per_sec": 0 00:27:49.378 }, 00:27:49.378 "claimed": false, 00:27:49.378 "zoned": false, 00:27:49.378 "supported_io_types": { 00:27:49.378 "read": true, 00:27:49.378 "write": true, 00:27:49.378 "unmap": true, 00:27:49.378 "flush": true, 00:27:49.378 "reset": true, 00:27:49.378 "nvme_admin": false, 00:27:49.378 "nvme_io": false, 00:27:49.378 "nvme_io_md": false, 00:27:49.378 "write_zeroes": true, 00:27:49.378 "zcopy": false, 00:27:49.378 "get_zone_info": false, 00:27:49.378 "zone_management": false, 00:27:49.378 "zone_append": false, 00:27:49.378 "compare": false, 00:27:49.378 "compare_and_write": false, 00:27:49.378 "abort": false, 00:27:49.378 "seek_hole": false, 00:27:49.378 "seek_data": false, 00:27:49.378 "copy": false, 00:27:49.378 "nvme_iov_md": false 00:27:49.378 }, 00:27:49.378 "memory_domains": [ 00:27:49.378 { 00:27:49.378 "dma_device_id": "system", 00:27:49.378 "dma_device_type": 1 00:27:49.378 }, 00:27:49.378 { 00:27:49.378 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:49.378 "dma_device_type": 2 00:27:49.378 }, 00:27:49.378 { 00:27:49.378 "dma_device_id": "system", 00:27:49.378 "dma_device_type": 1 00:27:49.378 }, 00:27:49.378 { 00:27:49.378 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:49.378 "dma_device_type": 2 00:27:49.378 }, 00:27:49.378 { 00:27:49.378 "dma_device_id": "system", 00:27:49.378 "dma_device_type": 1 00:27:49.378 }, 00:27:49.378 { 00:27:49.378 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:49.378 "dma_device_type": 2 00:27:49.378 } 00:27:49.378 ], 00:27:49.378 "driver_specific": { 00:27:49.378 "raid": { 00:27:49.378 "uuid": "bbb8aa17-176c-49cb-ade7-e8e1ff17ac37", 00:27:49.378 "strip_size_kb": 64, 00:27:49.378 "state": "online", 00:27:49.378 "raid_level": "raid0", 00:27:49.378 "superblock": true, 00:27:49.378 "num_base_bdevs": 3, 00:27:49.378 "num_base_bdevs_discovered": 3, 00:27:49.378 "num_base_bdevs_operational": 3, 00:27:49.378 "base_bdevs_list": [ 00:27:49.378 { 00:27:49.378 "name": "pt1", 00:27:49.378 "uuid": "00000000-0000-0000-0000-000000000001", 00:27:49.378 "is_configured": true, 00:27:49.378 "data_offset": 2048, 00:27:49.378 "data_size": 63488 00:27:49.378 }, 00:27:49.378 { 00:27:49.378 "name": "pt2", 00:27:49.378 "uuid": "00000000-0000-0000-0000-000000000002", 00:27:49.378 "is_configured": true, 00:27:49.378 "data_offset": 2048, 00:27:49.378 "data_size": 63488 00:27:49.378 }, 00:27:49.378 { 00:27:49.378 "name": "pt3", 00:27:49.378 "uuid": "00000000-0000-0000-0000-000000000003", 00:27:49.378 "is_configured": true, 00:27:49.378 "data_offset": 2048, 00:27:49.378 "data_size": 63488 00:27:49.378 } 00:27:49.378 ] 00:27:49.378 } 00:27:49.378 } 00:27:49.378 }' 00:27:49.378 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:27:49.637 pt2 00:27:49.637 pt3' 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:49.637 [2024-11-20 13:47:56.579677] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' bbb8aa17-176c-49cb-ade7-e8e1ff17ac37 '!=' bbb8aa17-176c-49cb-ade7-e8e1ff17ac37 ']' 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 63533 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 63533 ']' 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 63533 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 63533 00:27:49.637 killing process with pid 63533 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 63533' 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 63533 00:27:49.637 13:47:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 63533 00:27:49.637 [2024-11-20 13:47:56.629372] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:27:49.637 [2024-11-20 13:47:56.629495] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:49.637 [2024-11-20 13:47:56.629561] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:49.637 [2024-11-20 13:47:56.629575] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:27:49.895 [2024-11-20 13:47:56.828016] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:27:50.829 13:47:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:27:50.829 00:27:50.829 real 0m3.785s 00:27:50.829 user 0m5.328s 00:27:50.829 sys 0m0.641s 00:27:50.829 13:47:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:27:50.829 13:47:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:27:50.829 ************************************ 00:27:50.829 END TEST raid_superblock_test 00:27:50.829 ************************************ 00:27:50.829 13:47:57 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 3 read 00:27:50.829 13:47:57 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:27:50.829 13:47:57 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:27:50.829 13:47:57 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:27:50.829 ************************************ 00:27:50.829 START TEST raid_read_error_test 00:27:50.829 ************************************ 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 3 read 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:27:50.829 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.FLe0tukPsm 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=63770 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 63770 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 63770 ']' 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:50.829 13:47:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:50.829 [2024-11-20 13:47:57.650377] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:27:50.829 [2024-11-20 13:47:57.650505] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid63770 ] 00:27:50.829 [2024-11-20 13:47:57.801929] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:51.088 [2024-11-20 13:47:57.905487] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:27:51.088 [2024-11-20 13:47:58.032318] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:51.088 [2024-11-20 13:47:58.032372] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:51.655 BaseBdev1_malloc 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:51.655 true 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:51.655 [2024-11-20 13:47:58.565485] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:27:51.655 [2024-11-20 13:47:58.565555] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:51.655 [2024-11-20 13:47:58.565577] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:27:51.655 [2024-11-20 13:47:58.565588] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:51.655 [2024-11-20 13:47:58.567664] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:51.655 [2024-11-20 13:47:58.567846] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:27:51.655 BaseBdev1 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:51.655 BaseBdev2_malloc 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:51.655 true 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:51.655 [2024-11-20 13:47:58.607495] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:27:51.655 [2024-11-20 13:47:58.607552] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:51.655 [2024-11-20 13:47:58.607568] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:27:51.655 [2024-11-20 13:47:58.607577] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:51.655 [2024-11-20 13:47:58.609530] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:51.655 [2024-11-20 13:47:58.609567] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:27:51.655 BaseBdev2 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:51.655 BaseBdev3_malloc 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:51.655 true 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:51.655 [2024-11-20 13:47:58.666718] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:27:51.655 [2024-11-20 13:47:58.666910] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:51.655 [2024-11-20 13:47:58.666936] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:27:51.655 [2024-11-20 13:47:58.666947] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:51.655 [2024-11-20 13:47:58.668948] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:51.655 [2024-11-20 13:47:58.668991] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:27:51.655 BaseBdev3 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:51.655 [2024-11-20 13:47:58.674783] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:27:51.655 [2024-11-20 13:47:58.676561] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:27:51.655 [2024-11-20 13:47:58.676633] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:27:51.655 [2024-11-20 13:47:58.676809] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:27:51.655 [2024-11-20 13:47:58.676820] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:27:51.655 [2024-11-20 13:47:58.677085] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006560 00:27:51.655 [2024-11-20 13:47:58.677237] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:27:51.655 [2024-11-20 13:47:58.677291] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:27:51.655 [2024-11-20 13:47:58.677446] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:51.655 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:51.656 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:27:51.656 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:51.656 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:51.656 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:51.656 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:51.656 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:51.656 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:51.656 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:51.656 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:51.656 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:51.656 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:51.656 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:51.656 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:51.656 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:51.656 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:51.656 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:51.656 "name": "raid_bdev1", 00:27:51.656 "uuid": "77cacd49-0b88-4bb2-8bcd-674874e2dab9", 00:27:51.656 "strip_size_kb": 64, 00:27:51.656 "state": "online", 00:27:51.656 "raid_level": "raid0", 00:27:51.656 "superblock": true, 00:27:51.656 "num_base_bdevs": 3, 00:27:51.656 "num_base_bdevs_discovered": 3, 00:27:51.656 "num_base_bdevs_operational": 3, 00:27:51.656 "base_bdevs_list": [ 00:27:51.656 { 00:27:51.656 "name": "BaseBdev1", 00:27:51.656 "uuid": "f7f9401e-e7e8-5478-8750-70ccfe791e1d", 00:27:51.656 "is_configured": true, 00:27:51.656 "data_offset": 2048, 00:27:51.656 "data_size": 63488 00:27:51.656 }, 00:27:51.656 { 00:27:51.656 "name": "BaseBdev2", 00:27:51.656 "uuid": "28146349-f865-58f0-b0c3-7c94cd98f1d2", 00:27:51.656 "is_configured": true, 00:27:51.656 "data_offset": 2048, 00:27:51.656 "data_size": 63488 00:27:51.656 }, 00:27:51.656 { 00:27:51.656 "name": "BaseBdev3", 00:27:51.656 "uuid": "f5f47c9b-640d-5672-988e-1747f406fae5", 00:27:51.656 "is_configured": true, 00:27:51.656 "data_offset": 2048, 00:27:51.656 "data_size": 63488 00:27:51.656 } 00:27:51.656 ] 00:27:51.656 }' 00:27:51.656 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:51.656 13:47:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:52.219 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:27:52.219 13:47:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:27:52.219 [2024-11-20 13:47:59.087779] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006700 00:27:53.148 13:47:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:27:53.148 13:47:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:53.148 13:47:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:53.148 13:47:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:53.148 13:47:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:27:53.148 13:47:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:27:53.148 13:47:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:27:53.148 13:47:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:27:53.148 13:47:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:53.148 13:47:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:53.148 13:47:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:53.148 13:47:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:53.148 13:47:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:53.148 13:47:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:53.148 13:47:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:53.148 13:47:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:53.148 13:47:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:53.148 13:47:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:53.148 13:47:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:53.148 13:47:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:53.148 13:47:59 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:53.148 13:48:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:53.148 13:48:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:53.148 "name": "raid_bdev1", 00:27:53.148 "uuid": "77cacd49-0b88-4bb2-8bcd-674874e2dab9", 00:27:53.148 "strip_size_kb": 64, 00:27:53.148 "state": "online", 00:27:53.148 "raid_level": "raid0", 00:27:53.148 "superblock": true, 00:27:53.148 "num_base_bdevs": 3, 00:27:53.148 "num_base_bdevs_discovered": 3, 00:27:53.148 "num_base_bdevs_operational": 3, 00:27:53.148 "base_bdevs_list": [ 00:27:53.148 { 00:27:53.148 "name": "BaseBdev1", 00:27:53.148 "uuid": "f7f9401e-e7e8-5478-8750-70ccfe791e1d", 00:27:53.148 "is_configured": true, 00:27:53.148 "data_offset": 2048, 00:27:53.148 "data_size": 63488 00:27:53.148 }, 00:27:53.148 { 00:27:53.148 "name": "BaseBdev2", 00:27:53.148 "uuid": "28146349-f865-58f0-b0c3-7c94cd98f1d2", 00:27:53.148 "is_configured": true, 00:27:53.148 "data_offset": 2048, 00:27:53.148 "data_size": 63488 00:27:53.148 }, 00:27:53.148 { 00:27:53.148 "name": "BaseBdev3", 00:27:53.148 "uuid": "f5f47c9b-640d-5672-988e-1747f406fae5", 00:27:53.148 "is_configured": true, 00:27:53.148 "data_offset": 2048, 00:27:53.148 "data_size": 63488 00:27:53.148 } 00:27:53.148 ] 00:27:53.148 }' 00:27:53.148 13:48:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:53.148 13:48:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:53.404 13:48:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:27:53.404 13:48:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:53.404 13:48:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:53.404 [2024-11-20 13:48:00.353196] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:27:53.404 [2024-11-20 13:48:00.353230] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:27:53.404 [2024-11-20 13:48:00.355764] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:53.404 [2024-11-20 13:48:00.355808] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:53.404 [2024-11-20 13:48:00.355843] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:53.404 [2024-11-20 13:48:00.355852] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:27:53.404 { 00:27:53.404 "results": [ 00:27:53.404 { 00:27:53.404 "job": "raid_bdev1", 00:27:53.404 "core_mask": "0x1", 00:27:53.404 "workload": "randrw", 00:27:53.404 "percentage": 50, 00:27:53.404 "status": "finished", 00:27:53.404 "queue_depth": 1, 00:27:53.404 "io_size": 131072, 00:27:53.404 "runtime": 1.263764, 00:27:53.404 "iops": 15856.599808192035, 00:27:53.404 "mibps": 1982.0749760240044, 00:27:53.404 "io_failed": 1, 00:27:53.404 "io_timeout": 0, 00:27:53.404 "avg_latency_us": 86.93617810532781, 00:27:53.404 "min_latency_us": 26.584615384615386, 00:27:53.404 "max_latency_us": 1386.3384615384616 00:27:53.404 } 00:27:53.404 ], 00:27:53.404 "core_count": 1 00:27:53.404 } 00:27:53.404 13:48:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:53.404 13:48:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 63770 00:27:53.404 13:48:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 63770 ']' 00:27:53.404 13:48:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 63770 00:27:53.404 13:48:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:27:53.404 13:48:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:53.404 13:48:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 63770 00:27:53.404 killing process with pid 63770 00:27:53.404 13:48:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:27:53.404 13:48:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:27:53.404 13:48:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 63770' 00:27:53.404 13:48:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 63770 00:27:53.404 13:48:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 63770 00:27:53.404 [2024-11-20 13:48:00.385681] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:27:53.661 [2024-11-20 13:48:00.508753] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:27:54.235 13:48:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:27:54.235 13:48:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.FLe0tukPsm 00:27:54.235 13:48:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:27:54.235 13:48:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.79 00:27:54.235 13:48:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:27:54.235 13:48:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:27:54.235 13:48:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:27:54.235 13:48:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.79 != \0\.\0\0 ]] 00:27:54.235 00:27:54.235 real 0m3.594s 00:27:54.235 user 0m4.312s 00:27:54.235 sys 0m0.430s 00:27:54.235 ************************************ 00:27:54.235 END TEST raid_read_error_test 00:27:54.235 ************************************ 00:27:54.235 13:48:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:27:54.235 13:48:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:54.235 13:48:01 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 3 write 00:27:54.236 13:48:01 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:27:54.236 13:48:01 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:27:54.236 13:48:01 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:27:54.236 ************************************ 00:27:54.236 START TEST raid_write_error_test 00:27:54.236 ************************************ 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 3 write 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:27:54.236 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.p5GWpiWRTG 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=63904 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 63904 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 63904 ']' 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:27:54.236 13:48:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:54.494 [2024-11-20 13:48:01.294234] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:27:54.494 [2024-11-20 13:48:01.294511] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid63904 ] 00:27:54.494 [2024-11-20 13:48:01.453002] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:54.755 [2024-11-20 13:48:01.556787] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:27:54.755 [2024-11-20 13:48:01.683375] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:54.755 [2024-11-20 13:48:01.683451] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:55.321 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:55.321 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:27:55.321 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:27:55.321 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:27:55.321 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:55.321 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:55.322 BaseBdev1_malloc 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:55.322 true 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:55.322 [2024-11-20 13:48:02.196503] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:27:55.322 [2024-11-20 13:48:02.196558] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:55.322 [2024-11-20 13:48:02.196575] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:27:55.322 [2024-11-20 13:48:02.196585] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:55.322 [2024-11-20 13:48:02.198482] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:55.322 [2024-11-20 13:48:02.198621] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:27:55.322 BaseBdev1 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:55.322 BaseBdev2_malloc 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:55.322 true 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:55.322 [2024-11-20 13:48:02.242609] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:27:55.322 [2024-11-20 13:48:02.242662] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:55.322 [2024-11-20 13:48:02.242677] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:27:55.322 [2024-11-20 13:48:02.242687] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:55.322 [2024-11-20 13:48:02.244606] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:55.322 [2024-11-20 13:48:02.244638] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:27:55.322 BaseBdev2 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:55.322 BaseBdev3_malloc 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:55.322 true 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:55.322 [2024-11-20 13:48:02.304159] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:27:55.322 [2024-11-20 13:48:02.304211] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:55.322 [2024-11-20 13:48:02.304230] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:27:55.322 [2024-11-20 13:48:02.304240] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:55.322 [2024-11-20 13:48:02.306139] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:55.322 [2024-11-20 13:48:02.306169] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:27:55.322 BaseBdev3 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:55.322 [2024-11-20 13:48:02.312225] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:27:55.322 [2024-11-20 13:48:02.313834] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:27:55.322 [2024-11-20 13:48:02.314011] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:27:55.322 [2024-11-20 13:48:02.314186] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:27:55.322 [2024-11-20 13:48:02.314198] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:27:55.322 [2024-11-20 13:48:02.314418] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006560 00:27:55.322 [2024-11-20 13:48:02.314542] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:27:55.322 [2024-11-20 13:48:02.314553] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:27:55.322 [2024-11-20 13:48:02.314667] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:55.322 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:55.322 "name": "raid_bdev1", 00:27:55.322 "uuid": "1b2423e8-56e4-4290-9c5b-002a1966474a", 00:27:55.322 "strip_size_kb": 64, 00:27:55.322 "state": "online", 00:27:55.322 "raid_level": "raid0", 00:27:55.322 "superblock": true, 00:27:55.322 "num_base_bdevs": 3, 00:27:55.322 "num_base_bdevs_discovered": 3, 00:27:55.323 "num_base_bdevs_operational": 3, 00:27:55.323 "base_bdevs_list": [ 00:27:55.323 { 00:27:55.323 "name": "BaseBdev1", 00:27:55.323 "uuid": "cfa94ff0-1c0d-5ee9-ad6f-56ceb22cc8e8", 00:27:55.323 "is_configured": true, 00:27:55.323 "data_offset": 2048, 00:27:55.323 "data_size": 63488 00:27:55.323 }, 00:27:55.323 { 00:27:55.323 "name": "BaseBdev2", 00:27:55.323 "uuid": "ee7106c4-324b-50c8-8e7c-b50add97556c", 00:27:55.323 "is_configured": true, 00:27:55.323 "data_offset": 2048, 00:27:55.323 "data_size": 63488 00:27:55.323 }, 00:27:55.323 { 00:27:55.323 "name": "BaseBdev3", 00:27:55.323 "uuid": "0db4ae7b-682a-54fc-96ec-7db3c8ed346d", 00:27:55.323 "is_configured": true, 00:27:55.323 "data_offset": 2048, 00:27:55.323 "data_size": 63488 00:27:55.323 } 00:27:55.323 ] 00:27:55.323 }' 00:27:55.323 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:55.323 13:48:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:55.581 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:27:55.581 13:48:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:27:55.839 [2024-11-20 13:48:02.737190] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006700 00:27:56.774 13:48:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:27:56.774 13:48:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:56.774 13:48:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:56.774 13:48:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:56.774 13:48:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:27:56.774 13:48:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:27:56.774 13:48:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:27:56.774 13:48:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:27:56.775 13:48:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:56.775 13:48:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:56.775 13:48:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:27:56.775 13:48:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:56.775 13:48:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:56.775 13:48:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:56.775 13:48:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:56.775 13:48:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:56.775 13:48:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:56.775 13:48:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:56.775 13:48:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:56.775 13:48:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:56.775 13:48:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:56.775 13:48:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:56.775 13:48:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:56.775 "name": "raid_bdev1", 00:27:56.775 "uuid": "1b2423e8-56e4-4290-9c5b-002a1966474a", 00:27:56.775 "strip_size_kb": 64, 00:27:56.775 "state": "online", 00:27:56.775 "raid_level": "raid0", 00:27:56.775 "superblock": true, 00:27:56.775 "num_base_bdevs": 3, 00:27:56.775 "num_base_bdevs_discovered": 3, 00:27:56.775 "num_base_bdevs_operational": 3, 00:27:56.775 "base_bdevs_list": [ 00:27:56.775 { 00:27:56.775 "name": "BaseBdev1", 00:27:56.775 "uuid": "cfa94ff0-1c0d-5ee9-ad6f-56ceb22cc8e8", 00:27:56.775 "is_configured": true, 00:27:56.775 "data_offset": 2048, 00:27:56.775 "data_size": 63488 00:27:56.775 }, 00:27:56.775 { 00:27:56.775 "name": "BaseBdev2", 00:27:56.775 "uuid": "ee7106c4-324b-50c8-8e7c-b50add97556c", 00:27:56.775 "is_configured": true, 00:27:56.775 "data_offset": 2048, 00:27:56.775 "data_size": 63488 00:27:56.775 }, 00:27:56.775 { 00:27:56.775 "name": "BaseBdev3", 00:27:56.775 "uuid": "0db4ae7b-682a-54fc-96ec-7db3c8ed346d", 00:27:56.775 "is_configured": true, 00:27:56.775 "data_offset": 2048, 00:27:56.775 "data_size": 63488 00:27:56.775 } 00:27:56.775 ] 00:27:56.775 }' 00:27:56.775 13:48:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:56.775 13:48:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:57.033 13:48:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:27:57.033 13:48:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:57.033 13:48:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:57.033 [2024-11-20 13:48:03.947369] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:27:57.033 [2024-11-20 13:48:03.947547] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:27:57.033 [2024-11-20 13:48:03.950789] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:57.033 { 00:27:57.033 "results": [ 00:27:57.033 { 00:27:57.034 "job": "raid_bdev1", 00:27:57.034 "core_mask": "0x1", 00:27:57.034 "workload": "randrw", 00:27:57.034 "percentage": 50, 00:27:57.034 "status": "finished", 00:27:57.034 "queue_depth": 1, 00:27:57.034 "io_size": 131072, 00:27:57.034 "runtime": 1.20883, 00:27:57.034 "iops": 13577.591555470992, 00:27:57.034 "mibps": 1697.198944433874, 00:27:57.034 "io_failed": 1, 00:27:57.034 "io_timeout": 0, 00:27:57.034 "avg_latency_us": 101.02084861890881, 00:27:57.034 "min_latency_us": 25.993846153846153, 00:27:57.034 "max_latency_us": 1701.4153846153847 00:27:57.034 } 00:27:57.034 ], 00:27:57.034 "core_count": 1 00:27:57.034 } 00:27:57.034 [2024-11-20 13:48:03.950972] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:57.034 [2024-11-20 13:48:03.951036] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:57.034 [2024-11-20 13:48:03.951048] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:27:57.034 13:48:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:57.034 13:48:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 63904 00:27:57.034 13:48:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 63904 ']' 00:27:57.034 13:48:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 63904 00:27:57.034 13:48:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:27:57.034 13:48:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:57.034 13:48:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 63904 00:27:57.034 killing process with pid 63904 00:27:57.034 13:48:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:27:57.034 13:48:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:27:57.034 13:48:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 63904' 00:27:57.034 13:48:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 63904 00:27:57.034 13:48:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 63904 00:27:57.034 [2024-11-20 13:48:03.978617] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:27:57.292 [2024-11-20 13:48:04.131822] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:27:57.942 13:48:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.p5GWpiWRTG 00:27:57.942 13:48:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:27:57.942 13:48:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:27:57.942 ************************************ 00:27:57.942 END TEST raid_write_error_test 00:27:57.942 ************************************ 00:27:57.942 13:48:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.83 00:27:57.942 13:48:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:27:57.942 13:48:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:27:57.943 13:48:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:27:57.943 13:48:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.83 != \0\.\0\0 ]] 00:27:57.943 00:27:57.943 real 0m3.710s 00:27:57.943 user 0m4.413s 00:27:57.943 sys 0m0.423s 00:27:57.943 13:48:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:27:57.943 13:48:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:27:57.943 13:48:04 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:27:57.943 13:48:04 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 3 false 00:27:57.943 13:48:04 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:27:57.943 13:48:04 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:27:57.943 13:48:04 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:27:57.943 ************************************ 00:27:57.943 START TEST raid_state_function_test 00:27:57.943 ************************************ 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 3 false 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:27:57.943 Process raid pid: 64037 00:27:57.943 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=64037 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 64037' 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 64037 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 64037 ']' 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:57.943 13:48:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:58.201 [2024-11-20 13:48:05.047392] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:27:58.201 [2024-11-20 13:48:05.047538] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:27:58.201 [2024-11-20 13:48:05.202998] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:58.459 [2024-11-20 13:48:05.306603] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:27:58.459 [2024-11-20 13:48:05.430583] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:58.459 [2024-11-20 13:48:05.430631] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:59.025 13:48:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:59.025 13:48:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:27:59.025 13:48:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:27:59.025 13:48:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:59.025 13:48:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:59.025 [2024-11-20 13:48:05.909416] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:27:59.025 [2024-11-20 13:48:05.909470] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:27:59.025 [2024-11-20 13:48:05.909479] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:27:59.025 [2024-11-20 13:48:05.909487] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:27:59.025 [2024-11-20 13:48:05.909493] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:27:59.025 [2024-11-20 13:48:05.909500] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:27:59.025 13:48:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:59.025 13:48:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:27:59.025 13:48:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:59.025 13:48:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:59.025 13:48:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:27:59.025 13:48:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:59.025 13:48:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:59.025 13:48:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:59.025 13:48:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:59.025 13:48:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:59.025 13:48:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:59.025 13:48:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:59.025 13:48:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:59.025 13:48:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:59.025 13:48:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:59.025 13:48:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:59.025 13:48:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:59.025 "name": "Existed_Raid", 00:27:59.025 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:59.025 "strip_size_kb": 64, 00:27:59.025 "state": "configuring", 00:27:59.025 "raid_level": "concat", 00:27:59.025 "superblock": false, 00:27:59.025 "num_base_bdevs": 3, 00:27:59.025 "num_base_bdevs_discovered": 0, 00:27:59.025 "num_base_bdevs_operational": 3, 00:27:59.025 "base_bdevs_list": [ 00:27:59.025 { 00:27:59.025 "name": "BaseBdev1", 00:27:59.025 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:59.025 "is_configured": false, 00:27:59.025 "data_offset": 0, 00:27:59.025 "data_size": 0 00:27:59.025 }, 00:27:59.025 { 00:27:59.025 "name": "BaseBdev2", 00:27:59.025 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:59.025 "is_configured": false, 00:27:59.025 "data_offset": 0, 00:27:59.025 "data_size": 0 00:27:59.025 }, 00:27:59.025 { 00:27:59.025 "name": "BaseBdev3", 00:27:59.025 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:59.025 "is_configured": false, 00:27:59.025 "data_offset": 0, 00:27:59.025 "data_size": 0 00:27:59.025 } 00:27:59.025 ] 00:27:59.025 }' 00:27:59.025 13:48:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:59.025 13:48:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:59.284 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:27:59.284 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:59.284 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:59.284 [2024-11-20 13:48:06.241425] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:27:59.284 [2024-11-20 13:48:06.241460] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:27:59.284 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:59.284 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:27:59.284 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:59.285 [2024-11-20 13:48:06.249429] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:27:59.285 [2024-11-20 13:48:06.249470] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:27:59.285 [2024-11-20 13:48:06.249478] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:27:59.285 [2024-11-20 13:48:06.249485] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:27:59.285 [2024-11-20 13:48:06.249490] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:27:59.285 [2024-11-20 13:48:06.249497] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:59.285 [2024-11-20 13:48:06.279821] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:27:59.285 BaseBdev1 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:59.285 [ 00:27:59.285 { 00:27:59.285 "name": "BaseBdev1", 00:27:59.285 "aliases": [ 00:27:59.285 "06d7e28d-90f8-4fc2-b0a3-2fe140863f74" 00:27:59.285 ], 00:27:59.285 "product_name": "Malloc disk", 00:27:59.285 "block_size": 512, 00:27:59.285 "num_blocks": 65536, 00:27:59.285 "uuid": "06d7e28d-90f8-4fc2-b0a3-2fe140863f74", 00:27:59.285 "assigned_rate_limits": { 00:27:59.285 "rw_ios_per_sec": 0, 00:27:59.285 "rw_mbytes_per_sec": 0, 00:27:59.285 "r_mbytes_per_sec": 0, 00:27:59.285 "w_mbytes_per_sec": 0 00:27:59.285 }, 00:27:59.285 "claimed": true, 00:27:59.285 "claim_type": "exclusive_write", 00:27:59.285 "zoned": false, 00:27:59.285 "supported_io_types": { 00:27:59.285 "read": true, 00:27:59.285 "write": true, 00:27:59.285 "unmap": true, 00:27:59.285 "flush": true, 00:27:59.285 "reset": true, 00:27:59.285 "nvme_admin": false, 00:27:59.285 "nvme_io": false, 00:27:59.285 "nvme_io_md": false, 00:27:59.285 "write_zeroes": true, 00:27:59.285 "zcopy": true, 00:27:59.285 "get_zone_info": false, 00:27:59.285 "zone_management": false, 00:27:59.285 "zone_append": false, 00:27:59.285 "compare": false, 00:27:59.285 "compare_and_write": false, 00:27:59.285 "abort": true, 00:27:59.285 "seek_hole": false, 00:27:59.285 "seek_data": false, 00:27:59.285 "copy": true, 00:27:59.285 "nvme_iov_md": false 00:27:59.285 }, 00:27:59.285 "memory_domains": [ 00:27:59.285 { 00:27:59.285 "dma_device_id": "system", 00:27:59.285 "dma_device_type": 1 00:27:59.285 }, 00:27:59.285 { 00:27:59.285 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:27:59.285 "dma_device_type": 2 00:27:59.285 } 00:27:59.285 ], 00:27:59.285 "driver_specific": {} 00:27:59.285 } 00:27:59.285 ] 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:59.285 "name": "Existed_Raid", 00:27:59.285 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:59.285 "strip_size_kb": 64, 00:27:59.285 "state": "configuring", 00:27:59.285 "raid_level": "concat", 00:27:59.285 "superblock": false, 00:27:59.285 "num_base_bdevs": 3, 00:27:59.285 "num_base_bdevs_discovered": 1, 00:27:59.285 "num_base_bdevs_operational": 3, 00:27:59.285 "base_bdevs_list": [ 00:27:59.285 { 00:27:59.285 "name": "BaseBdev1", 00:27:59.285 "uuid": "06d7e28d-90f8-4fc2-b0a3-2fe140863f74", 00:27:59.285 "is_configured": true, 00:27:59.285 "data_offset": 0, 00:27:59.285 "data_size": 65536 00:27:59.285 }, 00:27:59.285 { 00:27:59.285 "name": "BaseBdev2", 00:27:59.285 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:59.285 "is_configured": false, 00:27:59.285 "data_offset": 0, 00:27:59.285 "data_size": 0 00:27:59.285 }, 00:27:59.285 { 00:27:59.285 "name": "BaseBdev3", 00:27:59.285 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:59.285 "is_configured": false, 00:27:59.285 "data_offset": 0, 00:27:59.285 "data_size": 0 00:27:59.285 } 00:27:59.285 ] 00:27:59.285 }' 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:59.285 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:59.544 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:27:59.544 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:59.544 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:59.803 [2024-11-20 13:48:06.603943] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:27:59.803 [2024-11-20 13:48:06.604142] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:27:59.803 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:59.803 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:27:59.803 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:59.803 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:59.803 [2024-11-20 13:48:06.612004] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:27:59.803 [2024-11-20 13:48:06.613766] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:27:59.803 [2024-11-20 13:48:06.613884] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:27:59.803 [2024-11-20 13:48:06.613935] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:27:59.803 [2024-11-20 13:48:06.613957] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:27:59.803 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:59.803 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:27:59.803 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:27:59.803 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:27:59.803 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:27:59.803 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:27:59.803 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:27:59.803 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:27:59.803 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:59.803 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:59.803 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:59.803 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:59.803 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:59.803 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:59.803 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:59.803 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:27:59.803 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:27:59.803 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:59.803 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:59.803 "name": "Existed_Raid", 00:27:59.803 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:59.803 "strip_size_kb": 64, 00:27:59.803 "state": "configuring", 00:27:59.803 "raid_level": "concat", 00:27:59.803 "superblock": false, 00:27:59.803 "num_base_bdevs": 3, 00:27:59.803 "num_base_bdevs_discovered": 1, 00:27:59.803 "num_base_bdevs_operational": 3, 00:27:59.803 "base_bdevs_list": [ 00:27:59.803 { 00:27:59.803 "name": "BaseBdev1", 00:27:59.803 "uuid": "06d7e28d-90f8-4fc2-b0a3-2fe140863f74", 00:27:59.803 "is_configured": true, 00:27:59.803 "data_offset": 0, 00:27:59.803 "data_size": 65536 00:27:59.803 }, 00:27:59.803 { 00:27:59.803 "name": "BaseBdev2", 00:27:59.803 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:59.803 "is_configured": false, 00:27:59.803 "data_offset": 0, 00:27:59.803 "data_size": 0 00:27:59.803 }, 00:27:59.803 { 00:27:59.803 "name": "BaseBdev3", 00:27:59.803 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:59.803 "is_configured": false, 00:27:59.803 "data_offset": 0, 00:27:59.803 "data_size": 0 00:27:59.803 } 00:27:59.803 ] 00:27:59.803 }' 00:27:59.803 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:59.803 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:00.062 [2024-11-20 13:48:06.969621] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:28:00.062 BaseBdev2 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:00.062 [ 00:28:00.062 { 00:28:00.062 "name": "BaseBdev2", 00:28:00.062 "aliases": [ 00:28:00.062 "d070e36c-7116-43f8-b186-5254e33173ec" 00:28:00.062 ], 00:28:00.062 "product_name": "Malloc disk", 00:28:00.062 "block_size": 512, 00:28:00.062 "num_blocks": 65536, 00:28:00.062 "uuid": "d070e36c-7116-43f8-b186-5254e33173ec", 00:28:00.062 "assigned_rate_limits": { 00:28:00.062 "rw_ios_per_sec": 0, 00:28:00.062 "rw_mbytes_per_sec": 0, 00:28:00.062 "r_mbytes_per_sec": 0, 00:28:00.062 "w_mbytes_per_sec": 0 00:28:00.062 }, 00:28:00.062 "claimed": true, 00:28:00.062 "claim_type": "exclusive_write", 00:28:00.062 "zoned": false, 00:28:00.062 "supported_io_types": { 00:28:00.062 "read": true, 00:28:00.062 "write": true, 00:28:00.062 "unmap": true, 00:28:00.062 "flush": true, 00:28:00.062 "reset": true, 00:28:00.062 "nvme_admin": false, 00:28:00.062 "nvme_io": false, 00:28:00.062 "nvme_io_md": false, 00:28:00.062 "write_zeroes": true, 00:28:00.062 "zcopy": true, 00:28:00.062 "get_zone_info": false, 00:28:00.062 "zone_management": false, 00:28:00.062 "zone_append": false, 00:28:00.062 "compare": false, 00:28:00.062 "compare_and_write": false, 00:28:00.062 "abort": true, 00:28:00.062 "seek_hole": false, 00:28:00.062 "seek_data": false, 00:28:00.062 "copy": true, 00:28:00.062 "nvme_iov_md": false 00:28:00.062 }, 00:28:00.062 "memory_domains": [ 00:28:00.062 { 00:28:00.062 "dma_device_id": "system", 00:28:00.062 "dma_device_type": 1 00:28:00.062 }, 00:28:00.062 { 00:28:00.062 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:00.062 "dma_device_type": 2 00:28:00.062 } 00:28:00.062 ], 00:28:00.062 "driver_specific": {} 00:28:00.062 } 00:28:00.062 ] 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:00.062 13:48:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:00.062 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:00.062 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:00.062 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:00.062 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:00.062 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:00.062 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:00.062 "name": "Existed_Raid", 00:28:00.062 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:00.062 "strip_size_kb": 64, 00:28:00.062 "state": "configuring", 00:28:00.062 "raid_level": "concat", 00:28:00.062 "superblock": false, 00:28:00.062 "num_base_bdevs": 3, 00:28:00.062 "num_base_bdevs_discovered": 2, 00:28:00.062 "num_base_bdevs_operational": 3, 00:28:00.062 "base_bdevs_list": [ 00:28:00.062 { 00:28:00.062 "name": "BaseBdev1", 00:28:00.062 "uuid": "06d7e28d-90f8-4fc2-b0a3-2fe140863f74", 00:28:00.062 "is_configured": true, 00:28:00.062 "data_offset": 0, 00:28:00.062 "data_size": 65536 00:28:00.062 }, 00:28:00.062 { 00:28:00.062 "name": "BaseBdev2", 00:28:00.062 "uuid": "d070e36c-7116-43f8-b186-5254e33173ec", 00:28:00.062 "is_configured": true, 00:28:00.062 "data_offset": 0, 00:28:00.062 "data_size": 65536 00:28:00.062 }, 00:28:00.062 { 00:28:00.062 "name": "BaseBdev3", 00:28:00.062 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:00.062 "is_configured": false, 00:28:00.062 "data_offset": 0, 00:28:00.062 "data_size": 0 00:28:00.062 } 00:28:00.062 ] 00:28:00.062 }' 00:28:00.062 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:00.062 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:00.321 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:28:00.321 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:00.321 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:00.321 [2024-11-20 13:48:07.351677] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:28:00.321 [2024-11-20 13:48:07.351729] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:28:00.321 [2024-11-20 13:48:07.351741] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:28:00.321 [2024-11-20 13:48:07.352036] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:28:00.321 [2024-11-20 13:48:07.352204] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:28:00.321 [2024-11-20 13:48:07.352214] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:28:00.321 [2024-11-20 13:48:07.352491] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:00.321 BaseBdev3 00:28:00.321 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:00.321 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:28:00.321 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:28:00.321 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:00.321 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:28:00.321 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:00.321 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:00.321 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:00.321 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:00.321 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:00.321 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:00.321 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:28:00.321 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:00.321 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:00.321 [ 00:28:00.321 { 00:28:00.321 "name": "BaseBdev3", 00:28:00.321 "aliases": [ 00:28:00.321 "83a8e028-c042-4cf3-b4f4-770c8712604b" 00:28:00.321 ], 00:28:00.321 "product_name": "Malloc disk", 00:28:00.321 "block_size": 512, 00:28:00.321 "num_blocks": 65536, 00:28:00.321 "uuid": "83a8e028-c042-4cf3-b4f4-770c8712604b", 00:28:00.321 "assigned_rate_limits": { 00:28:00.321 "rw_ios_per_sec": 0, 00:28:00.321 "rw_mbytes_per_sec": 0, 00:28:00.321 "r_mbytes_per_sec": 0, 00:28:00.321 "w_mbytes_per_sec": 0 00:28:00.321 }, 00:28:00.321 "claimed": true, 00:28:00.321 "claim_type": "exclusive_write", 00:28:00.321 "zoned": false, 00:28:00.321 "supported_io_types": { 00:28:00.321 "read": true, 00:28:00.321 "write": true, 00:28:00.321 "unmap": true, 00:28:00.321 "flush": true, 00:28:00.321 "reset": true, 00:28:00.321 "nvme_admin": false, 00:28:00.321 "nvme_io": false, 00:28:00.321 "nvme_io_md": false, 00:28:00.321 "write_zeroes": true, 00:28:00.321 "zcopy": true, 00:28:00.321 "get_zone_info": false, 00:28:00.321 "zone_management": false, 00:28:00.321 "zone_append": false, 00:28:00.321 "compare": false, 00:28:00.321 "compare_and_write": false, 00:28:00.321 "abort": true, 00:28:00.321 "seek_hole": false, 00:28:00.321 "seek_data": false, 00:28:00.321 "copy": true, 00:28:00.321 "nvme_iov_md": false 00:28:00.321 }, 00:28:00.321 "memory_domains": [ 00:28:00.321 { 00:28:00.321 "dma_device_id": "system", 00:28:00.321 "dma_device_type": 1 00:28:00.321 }, 00:28:00.321 { 00:28:00.321 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:00.321 "dma_device_type": 2 00:28:00.321 } 00:28:00.321 ], 00:28:00.321 "driver_specific": {} 00:28:00.321 } 00:28:00.321 ] 00:28:00.321 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:00.580 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:28:00.580 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:28:00.580 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:28:00.580 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:28:00.580 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:00.580 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:00.580 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:00.580 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:00.580 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:00.580 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:00.580 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:00.580 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:00.580 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:00.580 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:00.580 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:00.580 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:00.580 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:00.580 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:00.580 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:00.580 "name": "Existed_Raid", 00:28:00.580 "uuid": "b6390779-1a27-4a75-8fc4-5fcae4e61c9d", 00:28:00.580 "strip_size_kb": 64, 00:28:00.580 "state": "online", 00:28:00.580 "raid_level": "concat", 00:28:00.580 "superblock": false, 00:28:00.580 "num_base_bdevs": 3, 00:28:00.580 "num_base_bdevs_discovered": 3, 00:28:00.580 "num_base_bdevs_operational": 3, 00:28:00.580 "base_bdevs_list": [ 00:28:00.580 { 00:28:00.580 "name": "BaseBdev1", 00:28:00.580 "uuid": "06d7e28d-90f8-4fc2-b0a3-2fe140863f74", 00:28:00.580 "is_configured": true, 00:28:00.580 "data_offset": 0, 00:28:00.580 "data_size": 65536 00:28:00.580 }, 00:28:00.580 { 00:28:00.580 "name": "BaseBdev2", 00:28:00.580 "uuid": "d070e36c-7116-43f8-b186-5254e33173ec", 00:28:00.580 "is_configured": true, 00:28:00.580 "data_offset": 0, 00:28:00.580 "data_size": 65536 00:28:00.580 }, 00:28:00.580 { 00:28:00.580 "name": "BaseBdev3", 00:28:00.580 "uuid": "83a8e028-c042-4cf3-b4f4-770c8712604b", 00:28:00.580 "is_configured": true, 00:28:00.580 "data_offset": 0, 00:28:00.580 "data_size": 65536 00:28:00.580 } 00:28:00.580 ] 00:28:00.580 }' 00:28:00.580 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:00.580 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:00.839 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:28:00.839 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:28:00.839 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:28:00.839 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:28:00.839 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:28:00.839 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:28:00.839 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:28:00.839 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:28:00.839 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:00.839 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:00.839 [2024-11-20 13:48:07.676206] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:00.839 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:00.839 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:00.839 "name": "Existed_Raid", 00:28:00.839 "aliases": [ 00:28:00.839 "b6390779-1a27-4a75-8fc4-5fcae4e61c9d" 00:28:00.839 ], 00:28:00.839 "product_name": "Raid Volume", 00:28:00.839 "block_size": 512, 00:28:00.839 "num_blocks": 196608, 00:28:00.839 "uuid": "b6390779-1a27-4a75-8fc4-5fcae4e61c9d", 00:28:00.839 "assigned_rate_limits": { 00:28:00.839 "rw_ios_per_sec": 0, 00:28:00.839 "rw_mbytes_per_sec": 0, 00:28:00.839 "r_mbytes_per_sec": 0, 00:28:00.839 "w_mbytes_per_sec": 0 00:28:00.839 }, 00:28:00.839 "claimed": false, 00:28:00.839 "zoned": false, 00:28:00.839 "supported_io_types": { 00:28:00.839 "read": true, 00:28:00.839 "write": true, 00:28:00.839 "unmap": true, 00:28:00.839 "flush": true, 00:28:00.839 "reset": true, 00:28:00.839 "nvme_admin": false, 00:28:00.839 "nvme_io": false, 00:28:00.839 "nvme_io_md": false, 00:28:00.839 "write_zeroes": true, 00:28:00.839 "zcopy": false, 00:28:00.839 "get_zone_info": false, 00:28:00.839 "zone_management": false, 00:28:00.840 "zone_append": false, 00:28:00.840 "compare": false, 00:28:00.840 "compare_and_write": false, 00:28:00.840 "abort": false, 00:28:00.840 "seek_hole": false, 00:28:00.840 "seek_data": false, 00:28:00.840 "copy": false, 00:28:00.840 "nvme_iov_md": false 00:28:00.840 }, 00:28:00.840 "memory_domains": [ 00:28:00.840 { 00:28:00.840 "dma_device_id": "system", 00:28:00.840 "dma_device_type": 1 00:28:00.840 }, 00:28:00.840 { 00:28:00.840 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:00.840 "dma_device_type": 2 00:28:00.840 }, 00:28:00.840 { 00:28:00.840 "dma_device_id": "system", 00:28:00.840 "dma_device_type": 1 00:28:00.840 }, 00:28:00.840 { 00:28:00.840 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:00.840 "dma_device_type": 2 00:28:00.840 }, 00:28:00.840 { 00:28:00.840 "dma_device_id": "system", 00:28:00.840 "dma_device_type": 1 00:28:00.840 }, 00:28:00.840 { 00:28:00.840 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:00.840 "dma_device_type": 2 00:28:00.840 } 00:28:00.840 ], 00:28:00.840 "driver_specific": { 00:28:00.840 "raid": { 00:28:00.840 "uuid": "b6390779-1a27-4a75-8fc4-5fcae4e61c9d", 00:28:00.840 "strip_size_kb": 64, 00:28:00.840 "state": "online", 00:28:00.840 "raid_level": "concat", 00:28:00.840 "superblock": false, 00:28:00.840 "num_base_bdevs": 3, 00:28:00.840 "num_base_bdevs_discovered": 3, 00:28:00.840 "num_base_bdevs_operational": 3, 00:28:00.840 "base_bdevs_list": [ 00:28:00.840 { 00:28:00.840 "name": "BaseBdev1", 00:28:00.840 "uuid": "06d7e28d-90f8-4fc2-b0a3-2fe140863f74", 00:28:00.840 "is_configured": true, 00:28:00.840 "data_offset": 0, 00:28:00.840 "data_size": 65536 00:28:00.840 }, 00:28:00.840 { 00:28:00.840 "name": "BaseBdev2", 00:28:00.840 "uuid": "d070e36c-7116-43f8-b186-5254e33173ec", 00:28:00.840 "is_configured": true, 00:28:00.840 "data_offset": 0, 00:28:00.840 "data_size": 65536 00:28:00.840 }, 00:28:00.840 { 00:28:00.840 "name": "BaseBdev3", 00:28:00.840 "uuid": "83a8e028-c042-4cf3-b4f4-770c8712604b", 00:28:00.840 "is_configured": true, 00:28:00.840 "data_offset": 0, 00:28:00.840 "data_size": 65536 00:28:00.840 } 00:28:00.840 ] 00:28:00.840 } 00:28:00.840 } 00:28:00.840 }' 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:28:00.840 BaseBdev2 00:28:00.840 BaseBdev3' 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:00.840 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:00.840 [2024-11-20 13:48:07.855931] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:28:00.840 [2024-11-20 13:48:07.856116] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:28:00.840 [2024-11-20 13:48:07.856199] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:01.100 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:01.100 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:28:01.100 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:28:01.100 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:28:01.100 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:28:01.100 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:28:01.100 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 2 00:28:01.100 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:01.100 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:28:01.100 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:01.100 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:01.100 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:28:01.100 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:01.100 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:01.100 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:01.100 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:01.100 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:01.100 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:01.100 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:01.100 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:01.100 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:01.100 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:01.100 "name": "Existed_Raid", 00:28:01.100 "uuid": "b6390779-1a27-4a75-8fc4-5fcae4e61c9d", 00:28:01.100 "strip_size_kb": 64, 00:28:01.100 "state": "offline", 00:28:01.100 "raid_level": "concat", 00:28:01.100 "superblock": false, 00:28:01.100 "num_base_bdevs": 3, 00:28:01.100 "num_base_bdevs_discovered": 2, 00:28:01.100 "num_base_bdevs_operational": 2, 00:28:01.100 "base_bdevs_list": [ 00:28:01.100 { 00:28:01.100 "name": null, 00:28:01.100 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:01.100 "is_configured": false, 00:28:01.100 "data_offset": 0, 00:28:01.100 "data_size": 65536 00:28:01.100 }, 00:28:01.100 { 00:28:01.100 "name": "BaseBdev2", 00:28:01.100 "uuid": "d070e36c-7116-43f8-b186-5254e33173ec", 00:28:01.100 "is_configured": true, 00:28:01.100 "data_offset": 0, 00:28:01.100 "data_size": 65536 00:28:01.100 }, 00:28:01.100 { 00:28:01.100 "name": "BaseBdev3", 00:28:01.100 "uuid": "83a8e028-c042-4cf3-b4f4-770c8712604b", 00:28:01.100 "is_configured": true, 00:28:01.100 "data_offset": 0, 00:28:01.100 "data_size": 65536 00:28:01.100 } 00:28:01.100 ] 00:28:01.100 }' 00:28:01.100 13:48:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:01.100 13:48:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:01.359 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:28:01.359 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:28:01.359 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:01.359 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:01.359 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:28:01.359 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:01.359 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:01.359 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:28:01.359 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:28:01.359 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:28:01.359 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:01.359 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:01.359 [2024-11-20 13:48:08.281186] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:28:01.359 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:01.359 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:28:01.359 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:28:01.359 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:01.359 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:01.359 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:01.359 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:28:01.359 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:01.359 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:28:01.359 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:28:01.359 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:28:01.360 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:01.360 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:01.360 [2024-11-20 13:48:08.380468] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:28:01.360 [2024-11-20 13:48:08.380533] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:01.618 BaseBdev2 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:01.618 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:01.618 [ 00:28:01.618 { 00:28:01.618 "name": "BaseBdev2", 00:28:01.618 "aliases": [ 00:28:01.618 "281c9999-3dcb-401f-96b7-6986ee11c081" 00:28:01.618 ], 00:28:01.618 "product_name": "Malloc disk", 00:28:01.618 "block_size": 512, 00:28:01.618 "num_blocks": 65536, 00:28:01.618 "uuid": "281c9999-3dcb-401f-96b7-6986ee11c081", 00:28:01.618 "assigned_rate_limits": { 00:28:01.618 "rw_ios_per_sec": 0, 00:28:01.618 "rw_mbytes_per_sec": 0, 00:28:01.618 "r_mbytes_per_sec": 0, 00:28:01.618 "w_mbytes_per_sec": 0 00:28:01.618 }, 00:28:01.618 "claimed": false, 00:28:01.618 "zoned": false, 00:28:01.618 "supported_io_types": { 00:28:01.618 "read": true, 00:28:01.618 "write": true, 00:28:01.618 "unmap": true, 00:28:01.618 "flush": true, 00:28:01.618 "reset": true, 00:28:01.618 "nvme_admin": false, 00:28:01.618 "nvme_io": false, 00:28:01.618 "nvme_io_md": false, 00:28:01.618 "write_zeroes": true, 00:28:01.618 "zcopy": true, 00:28:01.618 "get_zone_info": false, 00:28:01.618 "zone_management": false, 00:28:01.619 "zone_append": false, 00:28:01.619 "compare": false, 00:28:01.619 "compare_and_write": false, 00:28:01.619 "abort": true, 00:28:01.619 "seek_hole": false, 00:28:01.619 "seek_data": false, 00:28:01.619 "copy": true, 00:28:01.619 "nvme_iov_md": false 00:28:01.619 }, 00:28:01.619 "memory_domains": [ 00:28:01.619 { 00:28:01.619 "dma_device_id": "system", 00:28:01.619 "dma_device_type": 1 00:28:01.619 }, 00:28:01.619 { 00:28:01.619 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:01.619 "dma_device_type": 2 00:28:01.619 } 00:28:01.619 ], 00:28:01.619 "driver_specific": {} 00:28:01.619 } 00:28:01.619 ] 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:01.619 BaseBdev3 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:01.619 [ 00:28:01.619 { 00:28:01.619 "name": "BaseBdev3", 00:28:01.619 "aliases": [ 00:28:01.619 "bfd6554e-7f3c-43b8-9600-113565492365" 00:28:01.619 ], 00:28:01.619 "product_name": "Malloc disk", 00:28:01.619 "block_size": 512, 00:28:01.619 "num_blocks": 65536, 00:28:01.619 "uuid": "bfd6554e-7f3c-43b8-9600-113565492365", 00:28:01.619 "assigned_rate_limits": { 00:28:01.619 "rw_ios_per_sec": 0, 00:28:01.619 "rw_mbytes_per_sec": 0, 00:28:01.619 "r_mbytes_per_sec": 0, 00:28:01.619 "w_mbytes_per_sec": 0 00:28:01.619 }, 00:28:01.619 "claimed": false, 00:28:01.619 "zoned": false, 00:28:01.619 "supported_io_types": { 00:28:01.619 "read": true, 00:28:01.619 "write": true, 00:28:01.619 "unmap": true, 00:28:01.619 "flush": true, 00:28:01.619 "reset": true, 00:28:01.619 "nvme_admin": false, 00:28:01.619 "nvme_io": false, 00:28:01.619 "nvme_io_md": false, 00:28:01.619 "write_zeroes": true, 00:28:01.619 "zcopy": true, 00:28:01.619 "get_zone_info": false, 00:28:01.619 "zone_management": false, 00:28:01.619 "zone_append": false, 00:28:01.619 "compare": false, 00:28:01.619 "compare_and_write": false, 00:28:01.619 "abort": true, 00:28:01.619 "seek_hole": false, 00:28:01.619 "seek_data": false, 00:28:01.619 "copy": true, 00:28:01.619 "nvme_iov_md": false 00:28:01.619 }, 00:28:01.619 "memory_domains": [ 00:28:01.619 { 00:28:01.619 "dma_device_id": "system", 00:28:01.619 "dma_device_type": 1 00:28:01.619 }, 00:28:01.619 { 00:28:01.619 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:01.619 "dma_device_type": 2 00:28:01.619 } 00:28:01.619 ], 00:28:01.619 "driver_specific": {} 00:28:01.619 } 00:28:01.619 ] 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:01.619 [2024-11-20 13:48:08.590746] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:28:01.619 [2024-11-20 13:48:08.590926] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:28:01.619 [2024-11-20 13:48:08.591023] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:28:01.619 [2024-11-20 13:48:08.593137] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:01.619 "name": "Existed_Raid", 00:28:01.619 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:01.619 "strip_size_kb": 64, 00:28:01.619 "state": "configuring", 00:28:01.619 "raid_level": "concat", 00:28:01.619 "superblock": false, 00:28:01.619 "num_base_bdevs": 3, 00:28:01.619 "num_base_bdevs_discovered": 2, 00:28:01.619 "num_base_bdevs_operational": 3, 00:28:01.619 "base_bdevs_list": [ 00:28:01.619 { 00:28:01.619 "name": "BaseBdev1", 00:28:01.619 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:01.619 "is_configured": false, 00:28:01.619 "data_offset": 0, 00:28:01.619 "data_size": 0 00:28:01.619 }, 00:28:01.619 { 00:28:01.619 "name": "BaseBdev2", 00:28:01.619 "uuid": "281c9999-3dcb-401f-96b7-6986ee11c081", 00:28:01.619 "is_configured": true, 00:28:01.619 "data_offset": 0, 00:28:01.619 "data_size": 65536 00:28:01.619 }, 00:28:01.619 { 00:28:01.619 "name": "BaseBdev3", 00:28:01.619 "uuid": "bfd6554e-7f3c-43b8-9600-113565492365", 00:28:01.619 "is_configured": true, 00:28:01.619 "data_offset": 0, 00:28:01.619 "data_size": 65536 00:28:01.619 } 00:28:01.619 ] 00:28:01.619 }' 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:01.619 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:01.879 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:28:01.879 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:01.879 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:01.879 [2024-11-20 13:48:08.906824] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:28:01.879 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:01.879 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:28:01.879 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:01.879 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:01.879 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:01.879 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:01.879 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:01.879 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:01.879 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:01.879 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:01.879 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:01.879 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:01.879 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:01.879 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:01.879 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:01.879 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:02.138 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:02.138 "name": "Existed_Raid", 00:28:02.138 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:02.138 "strip_size_kb": 64, 00:28:02.138 "state": "configuring", 00:28:02.138 "raid_level": "concat", 00:28:02.138 "superblock": false, 00:28:02.138 "num_base_bdevs": 3, 00:28:02.138 "num_base_bdevs_discovered": 1, 00:28:02.138 "num_base_bdevs_operational": 3, 00:28:02.138 "base_bdevs_list": [ 00:28:02.138 { 00:28:02.138 "name": "BaseBdev1", 00:28:02.138 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:02.138 "is_configured": false, 00:28:02.138 "data_offset": 0, 00:28:02.138 "data_size": 0 00:28:02.138 }, 00:28:02.138 { 00:28:02.138 "name": null, 00:28:02.138 "uuid": "281c9999-3dcb-401f-96b7-6986ee11c081", 00:28:02.138 "is_configured": false, 00:28:02.138 "data_offset": 0, 00:28:02.138 "data_size": 65536 00:28:02.138 }, 00:28:02.138 { 00:28:02.138 "name": "BaseBdev3", 00:28:02.138 "uuid": "bfd6554e-7f3c-43b8-9600-113565492365", 00:28:02.138 "is_configured": true, 00:28:02.138 "data_offset": 0, 00:28:02.138 "data_size": 65536 00:28:02.138 } 00:28:02.138 ] 00:28:02.138 }' 00:28:02.138 13:48:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:02.138 13:48:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:02.396 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:02.396 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:02.397 [2024-11-20 13:48:09.267370] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:02.397 BaseBdev1 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:02.397 [ 00:28:02.397 { 00:28:02.397 "name": "BaseBdev1", 00:28:02.397 "aliases": [ 00:28:02.397 "f403a3a8-7771-4eb3-a37d-00ecccc97a71" 00:28:02.397 ], 00:28:02.397 "product_name": "Malloc disk", 00:28:02.397 "block_size": 512, 00:28:02.397 "num_blocks": 65536, 00:28:02.397 "uuid": "f403a3a8-7771-4eb3-a37d-00ecccc97a71", 00:28:02.397 "assigned_rate_limits": { 00:28:02.397 "rw_ios_per_sec": 0, 00:28:02.397 "rw_mbytes_per_sec": 0, 00:28:02.397 "r_mbytes_per_sec": 0, 00:28:02.397 "w_mbytes_per_sec": 0 00:28:02.397 }, 00:28:02.397 "claimed": true, 00:28:02.397 "claim_type": "exclusive_write", 00:28:02.397 "zoned": false, 00:28:02.397 "supported_io_types": { 00:28:02.397 "read": true, 00:28:02.397 "write": true, 00:28:02.397 "unmap": true, 00:28:02.397 "flush": true, 00:28:02.397 "reset": true, 00:28:02.397 "nvme_admin": false, 00:28:02.397 "nvme_io": false, 00:28:02.397 "nvme_io_md": false, 00:28:02.397 "write_zeroes": true, 00:28:02.397 "zcopy": true, 00:28:02.397 "get_zone_info": false, 00:28:02.397 "zone_management": false, 00:28:02.397 "zone_append": false, 00:28:02.397 "compare": false, 00:28:02.397 "compare_and_write": false, 00:28:02.397 "abort": true, 00:28:02.397 "seek_hole": false, 00:28:02.397 "seek_data": false, 00:28:02.397 "copy": true, 00:28:02.397 "nvme_iov_md": false 00:28:02.397 }, 00:28:02.397 "memory_domains": [ 00:28:02.397 { 00:28:02.397 "dma_device_id": "system", 00:28:02.397 "dma_device_type": 1 00:28:02.397 }, 00:28:02.397 { 00:28:02.397 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:02.397 "dma_device_type": 2 00:28:02.397 } 00:28:02.397 ], 00:28:02.397 "driver_specific": {} 00:28:02.397 } 00:28:02.397 ] 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:02.397 "name": "Existed_Raid", 00:28:02.397 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:02.397 "strip_size_kb": 64, 00:28:02.397 "state": "configuring", 00:28:02.397 "raid_level": "concat", 00:28:02.397 "superblock": false, 00:28:02.397 "num_base_bdevs": 3, 00:28:02.397 "num_base_bdevs_discovered": 2, 00:28:02.397 "num_base_bdevs_operational": 3, 00:28:02.397 "base_bdevs_list": [ 00:28:02.397 { 00:28:02.397 "name": "BaseBdev1", 00:28:02.397 "uuid": "f403a3a8-7771-4eb3-a37d-00ecccc97a71", 00:28:02.397 "is_configured": true, 00:28:02.397 "data_offset": 0, 00:28:02.397 "data_size": 65536 00:28:02.397 }, 00:28:02.397 { 00:28:02.397 "name": null, 00:28:02.397 "uuid": "281c9999-3dcb-401f-96b7-6986ee11c081", 00:28:02.397 "is_configured": false, 00:28:02.397 "data_offset": 0, 00:28:02.397 "data_size": 65536 00:28:02.397 }, 00:28:02.397 { 00:28:02.397 "name": "BaseBdev3", 00:28:02.397 "uuid": "bfd6554e-7f3c-43b8-9600-113565492365", 00:28:02.397 "is_configured": true, 00:28:02.397 "data_offset": 0, 00:28:02.397 "data_size": 65536 00:28:02.397 } 00:28:02.397 ] 00:28:02.397 }' 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:02.397 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:02.656 [2024-11-20 13:48:09.643593] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:02.656 "name": "Existed_Raid", 00:28:02.656 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:02.656 "strip_size_kb": 64, 00:28:02.656 "state": "configuring", 00:28:02.656 "raid_level": "concat", 00:28:02.656 "superblock": false, 00:28:02.656 "num_base_bdevs": 3, 00:28:02.656 "num_base_bdevs_discovered": 1, 00:28:02.656 "num_base_bdevs_operational": 3, 00:28:02.656 "base_bdevs_list": [ 00:28:02.656 { 00:28:02.656 "name": "BaseBdev1", 00:28:02.656 "uuid": "f403a3a8-7771-4eb3-a37d-00ecccc97a71", 00:28:02.656 "is_configured": true, 00:28:02.656 "data_offset": 0, 00:28:02.656 "data_size": 65536 00:28:02.656 }, 00:28:02.656 { 00:28:02.656 "name": null, 00:28:02.656 "uuid": "281c9999-3dcb-401f-96b7-6986ee11c081", 00:28:02.656 "is_configured": false, 00:28:02.656 "data_offset": 0, 00:28:02.656 "data_size": 65536 00:28:02.656 }, 00:28:02.656 { 00:28:02.656 "name": null, 00:28:02.656 "uuid": "bfd6554e-7f3c-43b8-9600-113565492365", 00:28:02.656 "is_configured": false, 00:28:02.656 "data_offset": 0, 00:28:02.656 "data_size": 65536 00:28:02.656 } 00:28:02.656 ] 00:28:02.656 }' 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:02.656 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:02.914 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:28:02.914 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:02.914 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:02.914 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:02.914 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:03.173 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:28:03.173 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:28:03.173 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:03.173 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:03.173 [2024-11-20 13:48:09.983715] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:28:03.173 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:03.173 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:28:03.173 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:03.173 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:03.173 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:03.173 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:03.173 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:03.173 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:03.173 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:03.173 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:03.173 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:03.173 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:03.173 13:48:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:03.173 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:03.173 13:48:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:03.173 13:48:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:03.173 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:03.173 "name": "Existed_Raid", 00:28:03.173 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:03.173 "strip_size_kb": 64, 00:28:03.173 "state": "configuring", 00:28:03.173 "raid_level": "concat", 00:28:03.173 "superblock": false, 00:28:03.173 "num_base_bdevs": 3, 00:28:03.173 "num_base_bdevs_discovered": 2, 00:28:03.173 "num_base_bdevs_operational": 3, 00:28:03.173 "base_bdevs_list": [ 00:28:03.173 { 00:28:03.173 "name": "BaseBdev1", 00:28:03.173 "uuid": "f403a3a8-7771-4eb3-a37d-00ecccc97a71", 00:28:03.173 "is_configured": true, 00:28:03.173 "data_offset": 0, 00:28:03.173 "data_size": 65536 00:28:03.173 }, 00:28:03.173 { 00:28:03.173 "name": null, 00:28:03.173 "uuid": "281c9999-3dcb-401f-96b7-6986ee11c081", 00:28:03.173 "is_configured": false, 00:28:03.173 "data_offset": 0, 00:28:03.173 "data_size": 65536 00:28:03.173 }, 00:28:03.173 { 00:28:03.173 "name": "BaseBdev3", 00:28:03.173 "uuid": "bfd6554e-7f3c-43b8-9600-113565492365", 00:28:03.173 "is_configured": true, 00:28:03.173 "data_offset": 0, 00:28:03.173 "data_size": 65536 00:28:03.173 } 00:28:03.173 ] 00:28:03.173 }' 00:28:03.173 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:03.173 13:48:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:03.433 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:03.433 13:48:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:03.433 13:48:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:03.433 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:28:03.433 13:48:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:03.433 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:28:03.433 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:28:03.433 13:48:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:03.433 13:48:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:03.433 [2024-11-20 13:48:10.351802] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:28:03.433 13:48:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:03.433 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:28:03.433 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:03.433 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:03.433 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:03.433 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:03.433 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:03.433 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:03.433 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:03.433 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:03.433 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:03.433 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:03.433 13:48:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:03.433 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:03.433 13:48:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:03.433 13:48:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:03.433 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:03.433 "name": "Existed_Raid", 00:28:03.433 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:03.433 "strip_size_kb": 64, 00:28:03.433 "state": "configuring", 00:28:03.433 "raid_level": "concat", 00:28:03.433 "superblock": false, 00:28:03.433 "num_base_bdevs": 3, 00:28:03.433 "num_base_bdevs_discovered": 1, 00:28:03.433 "num_base_bdevs_operational": 3, 00:28:03.433 "base_bdevs_list": [ 00:28:03.433 { 00:28:03.433 "name": null, 00:28:03.434 "uuid": "f403a3a8-7771-4eb3-a37d-00ecccc97a71", 00:28:03.434 "is_configured": false, 00:28:03.434 "data_offset": 0, 00:28:03.434 "data_size": 65536 00:28:03.434 }, 00:28:03.434 { 00:28:03.434 "name": null, 00:28:03.434 "uuid": "281c9999-3dcb-401f-96b7-6986ee11c081", 00:28:03.434 "is_configured": false, 00:28:03.434 "data_offset": 0, 00:28:03.434 "data_size": 65536 00:28:03.434 }, 00:28:03.434 { 00:28:03.434 "name": "BaseBdev3", 00:28:03.434 "uuid": "bfd6554e-7f3c-43b8-9600-113565492365", 00:28:03.434 "is_configured": true, 00:28:03.434 "data_offset": 0, 00:28:03.434 "data_size": 65536 00:28:03.434 } 00:28:03.434 ] 00:28:03.434 }' 00:28:03.434 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:03.434 13:48:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:04.023 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:04.023 13:48:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:04.023 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:28:04.023 13:48:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:04.023 13:48:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:04.023 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:28:04.023 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:28:04.023 13:48:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:04.023 13:48:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:04.023 [2024-11-20 13:48:10.819855] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:28:04.024 13:48:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:04.024 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:28:04.024 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:04.024 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:04.024 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:04.024 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:04.024 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:04.024 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:04.024 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:04.024 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:04.024 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:04.024 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:04.024 13:48:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:04.024 13:48:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:04.024 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:04.024 13:48:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:04.024 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:04.024 "name": "Existed_Raid", 00:28:04.024 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:04.024 "strip_size_kb": 64, 00:28:04.024 "state": "configuring", 00:28:04.024 "raid_level": "concat", 00:28:04.024 "superblock": false, 00:28:04.024 "num_base_bdevs": 3, 00:28:04.024 "num_base_bdevs_discovered": 2, 00:28:04.024 "num_base_bdevs_operational": 3, 00:28:04.024 "base_bdevs_list": [ 00:28:04.024 { 00:28:04.024 "name": null, 00:28:04.024 "uuid": "f403a3a8-7771-4eb3-a37d-00ecccc97a71", 00:28:04.024 "is_configured": false, 00:28:04.024 "data_offset": 0, 00:28:04.024 "data_size": 65536 00:28:04.024 }, 00:28:04.024 { 00:28:04.024 "name": "BaseBdev2", 00:28:04.024 "uuid": "281c9999-3dcb-401f-96b7-6986ee11c081", 00:28:04.024 "is_configured": true, 00:28:04.024 "data_offset": 0, 00:28:04.024 "data_size": 65536 00:28:04.024 }, 00:28:04.024 { 00:28:04.024 "name": "BaseBdev3", 00:28:04.024 "uuid": "bfd6554e-7f3c-43b8-9600-113565492365", 00:28:04.024 "is_configured": true, 00:28:04.024 "data_offset": 0, 00:28:04.024 "data_size": 65536 00:28:04.024 } 00:28:04.024 ] 00:28:04.024 }' 00:28:04.024 13:48:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:04.024 13:48:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u f403a3a8-7771-4eb3-a37d-00ecccc97a71 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:04.282 [2024-11-20 13:48:11.214490] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:28:04.282 [2024-11-20 13:48:11.214571] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:28:04.282 [2024-11-20 13:48:11.214587] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:28:04.282 [2024-11-20 13:48:11.214868] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:28:04.282 [2024-11-20 13:48:11.215049] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:28:04.282 [2024-11-20 13:48:11.215106] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:28:04.282 [2024-11-20 13:48:11.215384] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:04.282 NewBaseBdev 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:04.282 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:04.282 [ 00:28:04.282 { 00:28:04.282 "name": "NewBaseBdev", 00:28:04.282 "aliases": [ 00:28:04.282 "f403a3a8-7771-4eb3-a37d-00ecccc97a71" 00:28:04.282 ], 00:28:04.282 "product_name": "Malloc disk", 00:28:04.282 "block_size": 512, 00:28:04.282 "num_blocks": 65536, 00:28:04.282 "uuid": "f403a3a8-7771-4eb3-a37d-00ecccc97a71", 00:28:04.282 "assigned_rate_limits": { 00:28:04.282 "rw_ios_per_sec": 0, 00:28:04.282 "rw_mbytes_per_sec": 0, 00:28:04.282 "r_mbytes_per_sec": 0, 00:28:04.282 "w_mbytes_per_sec": 0 00:28:04.282 }, 00:28:04.282 "claimed": true, 00:28:04.282 "claim_type": "exclusive_write", 00:28:04.282 "zoned": false, 00:28:04.282 "supported_io_types": { 00:28:04.282 "read": true, 00:28:04.282 "write": true, 00:28:04.282 "unmap": true, 00:28:04.282 "flush": true, 00:28:04.282 "reset": true, 00:28:04.282 "nvme_admin": false, 00:28:04.282 "nvme_io": false, 00:28:04.282 "nvme_io_md": false, 00:28:04.283 "write_zeroes": true, 00:28:04.283 "zcopy": true, 00:28:04.283 "get_zone_info": false, 00:28:04.283 "zone_management": false, 00:28:04.283 "zone_append": false, 00:28:04.283 "compare": false, 00:28:04.283 "compare_and_write": false, 00:28:04.283 "abort": true, 00:28:04.283 "seek_hole": false, 00:28:04.283 "seek_data": false, 00:28:04.283 "copy": true, 00:28:04.283 "nvme_iov_md": false 00:28:04.283 }, 00:28:04.283 "memory_domains": [ 00:28:04.283 { 00:28:04.283 "dma_device_id": "system", 00:28:04.283 "dma_device_type": 1 00:28:04.283 }, 00:28:04.283 { 00:28:04.283 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:04.283 "dma_device_type": 2 00:28:04.283 } 00:28:04.283 ], 00:28:04.283 "driver_specific": {} 00:28:04.283 } 00:28:04.283 ] 00:28:04.283 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:04.283 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:28:04.283 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:28:04.283 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:04.283 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:04.283 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:04.283 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:04.283 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:04.283 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:04.283 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:04.283 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:04.283 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:04.283 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:04.283 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:04.283 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:04.283 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:04.283 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:04.283 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:04.283 "name": "Existed_Raid", 00:28:04.283 "uuid": "cc28afa0-71a4-429e-8768-4b85ef47aed0", 00:28:04.283 "strip_size_kb": 64, 00:28:04.283 "state": "online", 00:28:04.283 "raid_level": "concat", 00:28:04.283 "superblock": false, 00:28:04.283 "num_base_bdevs": 3, 00:28:04.283 "num_base_bdevs_discovered": 3, 00:28:04.283 "num_base_bdevs_operational": 3, 00:28:04.283 "base_bdevs_list": [ 00:28:04.283 { 00:28:04.283 "name": "NewBaseBdev", 00:28:04.283 "uuid": "f403a3a8-7771-4eb3-a37d-00ecccc97a71", 00:28:04.283 "is_configured": true, 00:28:04.283 "data_offset": 0, 00:28:04.283 "data_size": 65536 00:28:04.283 }, 00:28:04.283 { 00:28:04.283 "name": "BaseBdev2", 00:28:04.283 "uuid": "281c9999-3dcb-401f-96b7-6986ee11c081", 00:28:04.283 "is_configured": true, 00:28:04.283 "data_offset": 0, 00:28:04.283 "data_size": 65536 00:28:04.283 }, 00:28:04.283 { 00:28:04.283 "name": "BaseBdev3", 00:28:04.283 "uuid": "bfd6554e-7f3c-43b8-9600-113565492365", 00:28:04.283 "is_configured": true, 00:28:04.283 "data_offset": 0, 00:28:04.283 "data_size": 65536 00:28:04.283 } 00:28:04.283 ] 00:28:04.283 }' 00:28:04.283 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:04.283 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:04.542 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:28:04.542 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:28:04.542 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:28:04.542 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:28:04.542 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:28:04.542 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:28:04.542 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:28:04.542 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:28:04.542 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:04.542 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:04.542 [2024-11-20 13:48:11.570974] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:04.542 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:04.542 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:04.542 "name": "Existed_Raid", 00:28:04.542 "aliases": [ 00:28:04.542 "cc28afa0-71a4-429e-8768-4b85ef47aed0" 00:28:04.542 ], 00:28:04.542 "product_name": "Raid Volume", 00:28:04.542 "block_size": 512, 00:28:04.542 "num_blocks": 196608, 00:28:04.542 "uuid": "cc28afa0-71a4-429e-8768-4b85ef47aed0", 00:28:04.542 "assigned_rate_limits": { 00:28:04.542 "rw_ios_per_sec": 0, 00:28:04.542 "rw_mbytes_per_sec": 0, 00:28:04.542 "r_mbytes_per_sec": 0, 00:28:04.542 "w_mbytes_per_sec": 0 00:28:04.542 }, 00:28:04.542 "claimed": false, 00:28:04.542 "zoned": false, 00:28:04.542 "supported_io_types": { 00:28:04.542 "read": true, 00:28:04.542 "write": true, 00:28:04.542 "unmap": true, 00:28:04.542 "flush": true, 00:28:04.542 "reset": true, 00:28:04.542 "nvme_admin": false, 00:28:04.542 "nvme_io": false, 00:28:04.542 "nvme_io_md": false, 00:28:04.542 "write_zeroes": true, 00:28:04.542 "zcopy": false, 00:28:04.542 "get_zone_info": false, 00:28:04.542 "zone_management": false, 00:28:04.542 "zone_append": false, 00:28:04.542 "compare": false, 00:28:04.542 "compare_and_write": false, 00:28:04.542 "abort": false, 00:28:04.542 "seek_hole": false, 00:28:04.542 "seek_data": false, 00:28:04.542 "copy": false, 00:28:04.542 "nvme_iov_md": false 00:28:04.542 }, 00:28:04.542 "memory_domains": [ 00:28:04.542 { 00:28:04.542 "dma_device_id": "system", 00:28:04.542 "dma_device_type": 1 00:28:04.542 }, 00:28:04.542 { 00:28:04.542 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:04.542 "dma_device_type": 2 00:28:04.542 }, 00:28:04.542 { 00:28:04.542 "dma_device_id": "system", 00:28:04.542 "dma_device_type": 1 00:28:04.542 }, 00:28:04.542 { 00:28:04.542 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:04.542 "dma_device_type": 2 00:28:04.542 }, 00:28:04.542 { 00:28:04.542 "dma_device_id": "system", 00:28:04.542 "dma_device_type": 1 00:28:04.542 }, 00:28:04.542 { 00:28:04.542 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:04.542 "dma_device_type": 2 00:28:04.542 } 00:28:04.542 ], 00:28:04.542 "driver_specific": { 00:28:04.542 "raid": { 00:28:04.542 "uuid": "cc28afa0-71a4-429e-8768-4b85ef47aed0", 00:28:04.542 "strip_size_kb": 64, 00:28:04.542 "state": "online", 00:28:04.542 "raid_level": "concat", 00:28:04.542 "superblock": false, 00:28:04.542 "num_base_bdevs": 3, 00:28:04.542 "num_base_bdevs_discovered": 3, 00:28:04.542 "num_base_bdevs_operational": 3, 00:28:04.542 "base_bdevs_list": [ 00:28:04.542 { 00:28:04.542 "name": "NewBaseBdev", 00:28:04.542 "uuid": "f403a3a8-7771-4eb3-a37d-00ecccc97a71", 00:28:04.542 "is_configured": true, 00:28:04.542 "data_offset": 0, 00:28:04.542 "data_size": 65536 00:28:04.542 }, 00:28:04.542 { 00:28:04.542 "name": "BaseBdev2", 00:28:04.542 "uuid": "281c9999-3dcb-401f-96b7-6986ee11c081", 00:28:04.542 "is_configured": true, 00:28:04.542 "data_offset": 0, 00:28:04.542 "data_size": 65536 00:28:04.542 }, 00:28:04.542 { 00:28:04.542 "name": "BaseBdev3", 00:28:04.542 "uuid": "bfd6554e-7f3c-43b8-9600-113565492365", 00:28:04.542 "is_configured": true, 00:28:04.542 "data_offset": 0, 00:28:04.542 "data_size": 65536 00:28:04.542 } 00:28:04.542 ] 00:28:04.542 } 00:28:04.542 } 00:28:04.542 }' 00:28:04.542 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:28:04.801 BaseBdev2 00:28:04.801 BaseBdev3' 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:04.801 [2024-11-20 13:48:11.738691] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:28:04.801 [2024-11-20 13:48:11.738734] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:28:04.801 [2024-11-20 13:48:11.738826] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:04.801 [2024-11-20 13:48:11.738898] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:04.801 [2024-11-20 13:48:11.738912] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 64037 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 64037 ']' 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 64037 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 64037 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:28:04.801 killing process with pid 64037 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 64037' 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 64037 00:28:04.801 [2024-11-20 13:48:11.769803] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:28:04.801 13:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 64037 00:28:05.059 [2024-11-20 13:48:11.974158] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:28:05.992 00:28:05.992 real 0m7.757s 00:28:05.992 user 0m12.200s 00:28:05.992 sys 0m1.349s 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:05.992 ************************************ 00:28:05.992 END TEST raid_state_function_test 00:28:05.992 ************************************ 00:28:05.992 13:48:12 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 3 true 00:28:05.992 13:48:12 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:28:05.992 13:48:12 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:28:05.992 13:48:12 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:28:05.992 ************************************ 00:28:05.992 START TEST raid_state_function_test_sb 00:28:05.992 ************************************ 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 3 true 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=64636 00:28:05.992 Process raid pid: 64636 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 64636' 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 64636 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 64636 ']' 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:28:05.992 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:28:05.992 13:48:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:28:05.993 13:48:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:28:05.993 13:48:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:05.993 13:48:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:28:05.993 [2024-11-20 13:48:12.842169] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:28:05.993 [2024-11-20 13:48:12.842290] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:28:05.993 [2024-11-20 13:48:13.000573] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:06.251 [2024-11-20 13:48:13.105561] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:28:06.251 [2024-11-20 13:48:13.230339] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:06.251 [2024-11-20 13:48:13.230391] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:06.818 13:48:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:28:06.818 13:48:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:28:06.818 13:48:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:28:06.818 13:48:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:06.818 13:48:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:06.818 [2024-11-20 13:48:13.705370] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:28:06.818 [2024-11-20 13:48:13.705439] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:28:06.818 [2024-11-20 13:48:13.705448] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:28:06.818 [2024-11-20 13:48:13.705456] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:28:06.818 [2024-11-20 13:48:13.705462] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:28:06.818 [2024-11-20 13:48:13.705469] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:28:06.818 13:48:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:06.818 13:48:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:28:06.818 13:48:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:06.818 13:48:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:06.818 13:48:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:06.818 13:48:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:06.818 13:48:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:06.818 13:48:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:06.818 13:48:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:06.818 13:48:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:06.818 13:48:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:06.818 13:48:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:06.818 13:48:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:06.818 13:48:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:06.818 13:48:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:06.818 13:48:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:06.818 13:48:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:06.818 "name": "Existed_Raid", 00:28:06.818 "uuid": "646eb8d9-bbbb-4347-ad0a-30135086617d", 00:28:06.818 "strip_size_kb": 64, 00:28:06.818 "state": "configuring", 00:28:06.818 "raid_level": "concat", 00:28:06.818 "superblock": true, 00:28:06.818 "num_base_bdevs": 3, 00:28:06.818 "num_base_bdevs_discovered": 0, 00:28:06.818 "num_base_bdevs_operational": 3, 00:28:06.818 "base_bdevs_list": [ 00:28:06.819 { 00:28:06.819 "name": "BaseBdev1", 00:28:06.819 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:06.819 "is_configured": false, 00:28:06.819 "data_offset": 0, 00:28:06.819 "data_size": 0 00:28:06.819 }, 00:28:06.819 { 00:28:06.819 "name": "BaseBdev2", 00:28:06.819 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:06.819 "is_configured": false, 00:28:06.819 "data_offset": 0, 00:28:06.819 "data_size": 0 00:28:06.819 }, 00:28:06.819 { 00:28:06.819 "name": "BaseBdev3", 00:28:06.819 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:06.819 "is_configured": false, 00:28:06.819 "data_offset": 0, 00:28:06.819 "data_size": 0 00:28:06.819 } 00:28:06.819 ] 00:28:06.819 }' 00:28:06.819 13:48:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:06.819 13:48:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:07.078 [2024-11-20 13:48:14.029350] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:28:07.078 [2024-11-20 13:48:14.029398] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:07.078 [2024-11-20 13:48:14.037360] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:28:07.078 [2024-11-20 13:48:14.037406] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:28:07.078 [2024-11-20 13:48:14.037414] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:28:07.078 [2024-11-20 13:48:14.037423] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:28:07.078 [2024-11-20 13:48:14.037428] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:28:07.078 [2024-11-20 13:48:14.037437] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:07.078 [2024-11-20 13:48:14.067776] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:07.078 BaseBdev1 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:07.078 [ 00:28:07.078 { 00:28:07.078 "name": "BaseBdev1", 00:28:07.078 "aliases": [ 00:28:07.078 "2011d552-37cf-4ad6-8574-c0bad56e588c" 00:28:07.078 ], 00:28:07.078 "product_name": "Malloc disk", 00:28:07.078 "block_size": 512, 00:28:07.078 "num_blocks": 65536, 00:28:07.078 "uuid": "2011d552-37cf-4ad6-8574-c0bad56e588c", 00:28:07.078 "assigned_rate_limits": { 00:28:07.078 "rw_ios_per_sec": 0, 00:28:07.078 "rw_mbytes_per_sec": 0, 00:28:07.078 "r_mbytes_per_sec": 0, 00:28:07.078 "w_mbytes_per_sec": 0 00:28:07.078 }, 00:28:07.078 "claimed": true, 00:28:07.078 "claim_type": "exclusive_write", 00:28:07.078 "zoned": false, 00:28:07.078 "supported_io_types": { 00:28:07.078 "read": true, 00:28:07.078 "write": true, 00:28:07.078 "unmap": true, 00:28:07.078 "flush": true, 00:28:07.078 "reset": true, 00:28:07.078 "nvme_admin": false, 00:28:07.078 "nvme_io": false, 00:28:07.078 "nvme_io_md": false, 00:28:07.078 "write_zeroes": true, 00:28:07.078 "zcopy": true, 00:28:07.078 "get_zone_info": false, 00:28:07.078 "zone_management": false, 00:28:07.078 "zone_append": false, 00:28:07.078 "compare": false, 00:28:07.078 "compare_and_write": false, 00:28:07.078 "abort": true, 00:28:07.078 "seek_hole": false, 00:28:07.078 "seek_data": false, 00:28:07.078 "copy": true, 00:28:07.078 "nvme_iov_md": false 00:28:07.078 }, 00:28:07.078 "memory_domains": [ 00:28:07.078 { 00:28:07.078 "dma_device_id": "system", 00:28:07.078 "dma_device_type": 1 00:28:07.078 }, 00:28:07.078 { 00:28:07.078 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:07.078 "dma_device_type": 2 00:28:07.078 } 00:28:07.078 ], 00:28:07.078 "driver_specific": {} 00:28:07.078 } 00:28:07.078 ] 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:07.078 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:07.078 "name": "Existed_Raid", 00:28:07.078 "uuid": "6dd06ed3-cd1d-44fa-98fb-a5bb2e6a57a2", 00:28:07.078 "strip_size_kb": 64, 00:28:07.079 "state": "configuring", 00:28:07.079 "raid_level": "concat", 00:28:07.079 "superblock": true, 00:28:07.079 "num_base_bdevs": 3, 00:28:07.079 "num_base_bdevs_discovered": 1, 00:28:07.079 "num_base_bdevs_operational": 3, 00:28:07.079 "base_bdevs_list": [ 00:28:07.079 { 00:28:07.079 "name": "BaseBdev1", 00:28:07.079 "uuid": "2011d552-37cf-4ad6-8574-c0bad56e588c", 00:28:07.079 "is_configured": true, 00:28:07.079 "data_offset": 2048, 00:28:07.079 "data_size": 63488 00:28:07.079 }, 00:28:07.079 { 00:28:07.079 "name": "BaseBdev2", 00:28:07.079 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:07.079 "is_configured": false, 00:28:07.079 "data_offset": 0, 00:28:07.079 "data_size": 0 00:28:07.079 }, 00:28:07.079 { 00:28:07.079 "name": "BaseBdev3", 00:28:07.079 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:07.079 "is_configured": false, 00:28:07.079 "data_offset": 0, 00:28:07.079 "data_size": 0 00:28:07.079 } 00:28:07.079 ] 00:28:07.079 }' 00:28:07.079 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:07.079 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:07.337 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:28:07.337 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:07.337 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:07.337 [2024-11-20 13:48:14.391890] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:28:07.337 [2024-11-20 13:48:14.391952] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:28:07.595 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:07.595 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:28:07.595 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:07.595 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:07.595 [2024-11-20 13:48:14.399952] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:07.595 [2024-11-20 13:48:14.401626] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:28:07.595 [2024-11-20 13:48:14.401671] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:28:07.596 [2024-11-20 13:48:14.401680] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:28:07.596 [2024-11-20 13:48:14.401688] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:28:07.596 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:07.596 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:28:07.596 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:28:07.596 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:28:07.596 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:07.596 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:07.596 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:07.596 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:07.596 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:07.596 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:07.596 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:07.596 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:07.596 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:07.596 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:07.596 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:07.596 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:07.596 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:07.596 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:07.596 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:07.596 "name": "Existed_Raid", 00:28:07.596 "uuid": "dd917ec7-af85-48c3-b361-6c2c3cb3fc41", 00:28:07.596 "strip_size_kb": 64, 00:28:07.596 "state": "configuring", 00:28:07.596 "raid_level": "concat", 00:28:07.596 "superblock": true, 00:28:07.596 "num_base_bdevs": 3, 00:28:07.596 "num_base_bdevs_discovered": 1, 00:28:07.596 "num_base_bdevs_operational": 3, 00:28:07.596 "base_bdevs_list": [ 00:28:07.596 { 00:28:07.596 "name": "BaseBdev1", 00:28:07.596 "uuid": "2011d552-37cf-4ad6-8574-c0bad56e588c", 00:28:07.596 "is_configured": true, 00:28:07.596 "data_offset": 2048, 00:28:07.596 "data_size": 63488 00:28:07.596 }, 00:28:07.596 { 00:28:07.596 "name": "BaseBdev2", 00:28:07.596 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:07.596 "is_configured": false, 00:28:07.596 "data_offset": 0, 00:28:07.596 "data_size": 0 00:28:07.596 }, 00:28:07.596 { 00:28:07.596 "name": "BaseBdev3", 00:28:07.596 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:07.596 "is_configured": false, 00:28:07.596 "data_offset": 0, 00:28:07.596 "data_size": 0 00:28:07.596 } 00:28:07.596 ] 00:28:07.596 }' 00:28:07.596 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:07.596 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:07.855 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:28:07.855 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:07.855 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:07.855 [2024-11-20 13:48:14.748669] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:28:07.855 BaseBdev2 00:28:07.855 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:07.855 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:28:07.855 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:28:07.855 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:07.855 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:28:07.855 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:07.855 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:07.855 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:07.855 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:07.855 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:07.856 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:07.856 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:28:07.856 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:07.856 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:07.856 [ 00:28:07.856 { 00:28:07.856 "name": "BaseBdev2", 00:28:07.856 "aliases": [ 00:28:07.856 "e97c090c-8817-4644-b8c6-b4d9d220dd7e" 00:28:07.856 ], 00:28:07.856 "product_name": "Malloc disk", 00:28:07.856 "block_size": 512, 00:28:07.856 "num_blocks": 65536, 00:28:07.856 "uuid": "e97c090c-8817-4644-b8c6-b4d9d220dd7e", 00:28:07.856 "assigned_rate_limits": { 00:28:07.856 "rw_ios_per_sec": 0, 00:28:07.856 "rw_mbytes_per_sec": 0, 00:28:07.856 "r_mbytes_per_sec": 0, 00:28:07.856 "w_mbytes_per_sec": 0 00:28:07.856 }, 00:28:07.856 "claimed": true, 00:28:07.856 "claim_type": "exclusive_write", 00:28:07.856 "zoned": false, 00:28:07.856 "supported_io_types": { 00:28:07.856 "read": true, 00:28:07.856 "write": true, 00:28:07.856 "unmap": true, 00:28:07.856 "flush": true, 00:28:07.856 "reset": true, 00:28:07.856 "nvme_admin": false, 00:28:07.856 "nvme_io": false, 00:28:07.856 "nvme_io_md": false, 00:28:07.856 "write_zeroes": true, 00:28:07.856 "zcopy": true, 00:28:07.856 "get_zone_info": false, 00:28:07.856 "zone_management": false, 00:28:07.856 "zone_append": false, 00:28:07.856 "compare": false, 00:28:07.856 "compare_and_write": false, 00:28:07.856 "abort": true, 00:28:07.856 "seek_hole": false, 00:28:07.856 "seek_data": false, 00:28:07.856 "copy": true, 00:28:07.856 "nvme_iov_md": false 00:28:07.856 }, 00:28:07.856 "memory_domains": [ 00:28:07.856 { 00:28:07.856 "dma_device_id": "system", 00:28:07.856 "dma_device_type": 1 00:28:07.856 }, 00:28:07.856 { 00:28:07.856 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:07.856 "dma_device_type": 2 00:28:07.856 } 00:28:07.856 ], 00:28:07.856 "driver_specific": {} 00:28:07.856 } 00:28:07.856 ] 00:28:07.856 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:07.856 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:28:07.856 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:28:07.856 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:28:07.856 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:28:07.856 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:07.856 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:07.856 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:07.856 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:07.856 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:07.856 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:07.856 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:07.856 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:07.856 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:07.856 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:07.856 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:07.856 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:07.856 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:07.856 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:07.856 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:07.856 "name": "Existed_Raid", 00:28:07.856 "uuid": "dd917ec7-af85-48c3-b361-6c2c3cb3fc41", 00:28:07.856 "strip_size_kb": 64, 00:28:07.856 "state": "configuring", 00:28:07.856 "raid_level": "concat", 00:28:07.856 "superblock": true, 00:28:07.856 "num_base_bdevs": 3, 00:28:07.856 "num_base_bdevs_discovered": 2, 00:28:07.856 "num_base_bdevs_operational": 3, 00:28:07.856 "base_bdevs_list": [ 00:28:07.856 { 00:28:07.856 "name": "BaseBdev1", 00:28:07.856 "uuid": "2011d552-37cf-4ad6-8574-c0bad56e588c", 00:28:07.856 "is_configured": true, 00:28:07.856 "data_offset": 2048, 00:28:07.856 "data_size": 63488 00:28:07.856 }, 00:28:07.856 { 00:28:07.856 "name": "BaseBdev2", 00:28:07.856 "uuid": "e97c090c-8817-4644-b8c6-b4d9d220dd7e", 00:28:07.856 "is_configured": true, 00:28:07.856 "data_offset": 2048, 00:28:07.856 "data_size": 63488 00:28:07.856 }, 00:28:07.856 { 00:28:07.856 "name": "BaseBdev3", 00:28:07.856 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:07.856 "is_configured": false, 00:28:07.856 "data_offset": 0, 00:28:07.856 "data_size": 0 00:28:07.856 } 00:28:07.856 ] 00:28:07.856 }' 00:28:07.856 13:48:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:07.856 13:48:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:08.114 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:28:08.114 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:08.114 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:08.114 [2024-11-20 13:48:15.133716] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:28:08.114 [2024-11-20 13:48:15.133974] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:28:08.114 [2024-11-20 13:48:15.134016] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:28:08.114 [2024-11-20 13:48:15.134248] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:28:08.114 [2024-11-20 13:48:15.134385] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:28:08.114 [2024-11-20 13:48:15.134399] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:28:08.114 BaseBdev3 00:28:08.114 [2024-11-20 13:48:15.134518] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:08.114 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:08.114 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:28:08.114 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:28:08.114 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:08.114 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:28:08.114 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:08.114 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:08.114 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:08.114 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:08.114 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:08.114 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:08.114 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:28:08.114 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:08.114 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:08.114 [ 00:28:08.114 { 00:28:08.115 "name": "BaseBdev3", 00:28:08.115 "aliases": [ 00:28:08.115 "3ee71256-f087-4de3-a694-059d3ebd3d37" 00:28:08.115 ], 00:28:08.115 "product_name": "Malloc disk", 00:28:08.115 "block_size": 512, 00:28:08.115 "num_blocks": 65536, 00:28:08.115 "uuid": "3ee71256-f087-4de3-a694-059d3ebd3d37", 00:28:08.115 "assigned_rate_limits": { 00:28:08.115 "rw_ios_per_sec": 0, 00:28:08.115 "rw_mbytes_per_sec": 0, 00:28:08.115 "r_mbytes_per_sec": 0, 00:28:08.115 "w_mbytes_per_sec": 0 00:28:08.115 }, 00:28:08.115 "claimed": true, 00:28:08.115 "claim_type": "exclusive_write", 00:28:08.115 "zoned": false, 00:28:08.115 "supported_io_types": { 00:28:08.115 "read": true, 00:28:08.115 "write": true, 00:28:08.115 "unmap": true, 00:28:08.115 "flush": true, 00:28:08.115 "reset": true, 00:28:08.115 "nvme_admin": false, 00:28:08.115 "nvme_io": false, 00:28:08.115 "nvme_io_md": false, 00:28:08.115 "write_zeroes": true, 00:28:08.115 "zcopy": true, 00:28:08.115 "get_zone_info": false, 00:28:08.115 "zone_management": false, 00:28:08.115 "zone_append": false, 00:28:08.115 "compare": false, 00:28:08.115 "compare_and_write": false, 00:28:08.115 "abort": true, 00:28:08.115 "seek_hole": false, 00:28:08.115 "seek_data": false, 00:28:08.115 "copy": true, 00:28:08.115 "nvme_iov_md": false 00:28:08.115 }, 00:28:08.115 "memory_domains": [ 00:28:08.115 { 00:28:08.115 "dma_device_id": "system", 00:28:08.115 "dma_device_type": 1 00:28:08.115 }, 00:28:08.115 { 00:28:08.115 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:08.115 "dma_device_type": 2 00:28:08.115 } 00:28:08.115 ], 00:28:08.115 "driver_specific": {} 00:28:08.115 } 00:28:08.115 ] 00:28:08.115 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:08.115 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:28:08.115 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:28:08.115 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:28:08.115 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:28:08.115 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:08.115 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:08.115 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:08.115 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:08.115 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:08.115 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:08.115 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:08.115 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:08.115 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:08.115 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:08.115 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:08.115 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:08.115 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:08.373 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:08.373 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:08.373 "name": "Existed_Raid", 00:28:08.373 "uuid": "dd917ec7-af85-48c3-b361-6c2c3cb3fc41", 00:28:08.373 "strip_size_kb": 64, 00:28:08.373 "state": "online", 00:28:08.373 "raid_level": "concat", 00:28:08.373 "superblock": true, 00:28:08.373 "num_base_bdevs": 3, 00:28:08.373 "num_base_bdevs_discovered": 3, 00:28:08.373 "num_base_bdevs_operational": 3, 00:28:08.373 "base_bdevs_list": [ 00:28:08.373 { 00:28:08.373 "name": "BaseBdev1", 00:28:08.373 "uuid": "2011d552-37cf-4ad6-8574-c0bad56e588c", 00:28:08.373 "is_configured": true, 00:28:08.373 "data_offset": 2048, 00:28:08.373 "data_size": 63488 00:28:08.373 }, 00:28:08.373 { 00:28:08.373 "name": "BaseBdev2", 00:28:08.373 "uuid": "e97c090c-8817-4644-b8c6-b4d9d220dd7e", 00:28:08.373 "is_configured": true, 00:28:08.373 "data_offset": 2048, 00:28:08.373 "data_size": 63488 00:28:08.373 }, 00:28:08.373 { 00:28:08.373 "name": "BaseBdev3", 00:28:08.373 "uuid": "3ee71256-f087-4de3-a694-059d3ebd3d37", 00:28:08.373 "is_configured": true, 00:28:08.373 "data_offset": 2048, 00:28:08.373 "data_size": 63488 00:28:08.373 } 00:28:08.373 ] 00:28:08.373 }' 00:28:08.373 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:08.373 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:08.632 [2024-11-20 13:48:15.498175] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:08.632 "name": "Existed_Raid", 00:28:08.632 "aliases": [ 00:28:08.632 "dd917ec7-af85-48c3-b361-6c2c3cb3fc41" 00:28:08.632 ], 00:28:08.632 "product_name": "Raid Volume", 00:28:08.632 "block_size": 512, 00:28:08.632 "num_blocks": 190464, 00:28:08.632 "uuid": "dd917ec7-af85-48c3-b361-6c2c3cb3fc41", 00:28:08.632 "assigned_rate_limits": { 00:28:08.632 "rw_ios_per_sec": 0, 00:28:08.632 "rw_mbytes_per_sec": 0, 00:28:08.632 "r_mbytes_per_sec": 0, 00:28:08.632 "w_mbytes_per_sec": 0 00:28:08.632 }, 00:28:08.632 "claimed": false, 00:28:08.632 "zoned": false, 00:28:08.632 "supported_io_types": { 00:28:08.632 "read": true, 00:28:08.632 "write": true, 00:28:08.632 "unmap": true, 00:28:08.632 "flush": true, 00:28:08.632 "reset": true, 00:28:08.632 "nvme_admin": false, 00:28:08.632 "nvme_io": false, 00:28:08.632 "nvme_io_md": false, 00:28:08.632 "write_zeroes": true, 00:28:08.632 "zcopy": false, 00:28:08.632 "get_zone_info": false, 00:28:08.632 "zone_management": false, 00:28:08.632 "zone_append": false, 00:28:08.632 "compare": false, 00:28:08.632 "compare_and_write": false, 00:28:08.632 "abort": false, 00:28:08.632 "seek_hole": false, 00:28:08.632 "seek_data": false, 00:28:08.632 "copy": false, 00:28:08.632 "nvme_iov_md": false 00:28:08.632 }, 00:28:08.632 "memory_domains": [ 00:28:08.632 { 00:28:08.632 "dma_device_id": "system", 00:28:08.632 "dma_device_type": 1 00:28:08.632 }, 00:28:08.632 { 00:28:08.632 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:08.632 "dma_device_type": 2 00:28:08.632 }, 00:28:08.632 { 00:28:08.632 "dma_device_id": "system", 00:28:08.632 "dma_device_type": 1 00:28:08.632 }, 00:28:08.632 { 00:28:08.632 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:08.632 "dma_device_type": 2 00:28:08.632 }, 00:28:08.632 { 00:28:08.632 "dma_device_id": "system", 00:28:08.632 "dma_device_type": 1 00:28:08.632 }, 00:28:08.632 { 00:28:08.632 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:08.632 "dma_device_type": 2 00:28:08.632 } 00:28:08.632 ], 00:28:08.632 "driver_specific": { 00:28:08.632 "raid": { 00:28:08.632 "uuid": "dd917ec7-af85-48c3-b361-6c2c3cb3fc41", 00:28:08.632 "strip_size_kb": 64, 00:28:08.632 "state": "online", 00:28:08.632 "raid_level": "concat", 00:28:08.632 "superblock": true, 00:28:08.632 "num_base_bdevs": 3, 00:28:08.632 "num_base_bdevs_discovered": 3, 00:28:08.632 "num_base_bdevs_operational": 3, 00:28:08.632 "base_bdevs_list": [ 00:28:08.632 { 00:28:08.632 "name": "BaseBdev1", 00:28:08.632 "uuid": "2011d552-37cf-4ad6-8574-c0bad56e588c", 00:28:08.632 "is_configured": true, 00:28:08.632 "data_offset": 2048, 00:28:08.632 "data_size": 63488 00:28:08.632 }, 00:28:08.632 { 00:28:08.632 "name": "BaseBdev2", 00:28:08.632 "uuid": "e97c090c-8817-4644-b8c6-b4d9d220dd7e", 00:28:08.632 "is_configured": true, 00:28:08.632 "data_offset": 2048, 00:28:08.632 "data_size": 63488 00:28:08.632 }, 00:28:08.632 { 00:28:08.632 "name": "BaseBdev3", 00:28:08.632 "uuid": "3ee71256-f087-4de3-a694-059d3ebd3d37", 00:28:08.632 "is_configured": true, 00:28:08.632 "data_offset": 2048, 00:28:08.632 "data_size": 63488 00:28:08.632 } 00:28:08.632 ] 00:28:08.632 } 00:28:08.632 } 00:28:08.632 }' 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:28:08.632 BaseBdev2 00:28:08.632 BaseBdev3' 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:08.632 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:08.890 [2024-11-20 13:48:15.717940] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:28:08.890 [2024-11-20 13:48:15.717996] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:28:08.890 [2024-11-20 13:48:15.718051] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 2 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:08.890 "name": "Existed_Raid", 00:28:08.890 "uuid": "dd917ec7-af85-48c3-b361-6c2c3cb3fc41", 00:28:08.890 "strip_size_kb": 64, 00:28:08.890 "state": "offline", 00:28:08.890 "raid_level": "concat", 00:28:08.890 "superblock": true, 00:28:08.890 "num_base_bdevs": 3, 00:28:08.890 "num_base_bdevs_discovered": 2, 00:28:08.890 "num_base_bdevs_operational": 2, 00:28:08.890 "base_bdevs_list": [ 00:28:08.890 { 00:28:08.890 "name": null, 00:28:08.890 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:08.890 "is_configured": false, 00:28:08.890 "data_offset": 0, 00:28:08.890 "data_size": 63488 00:28:08.890 }, 00:28:08.890 { 00:28:08.890 "name": "BaseBdev2", 00:28:08.890 "uuid": "e97c090c-8817-4644-b8c6-b4d9d220dd7e", 00:28:08.890 "is_configured": true, 00:28:08.890 "data_offset": 2048, 00:28:08.890 "data_size": 63488 00:28:08.890 }, 00:28:08.890 { 00:28:08.890 "name": "BaseBdev3", 00:28:08.890 "uuid": "3ee71256-f087-4de3-a694-059d3ebd3d37", 00:28:08.890 "is_configured": true, 00:28:08.890 "data_offset": 2048, 00:28:08.890 "data_size": 63488 00:28:08.890 } 00:28:08.890 ] 00:28:08.890 }' 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:08.890 13:48:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:09.149 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:28:09.149 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:28:09.149 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:28:09.149 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:09.149 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:09.149 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:09.149 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:09.149 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:28:09.149 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:28:09.149 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:28:09.149 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:09.149 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:09.149 [2024-11-20 13:48:16.093152] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:28:09.149 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:09.149 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:28:09.149 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:28:09.149 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:09.149 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:28:09.149 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:09.149 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:09.149 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:09.149 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:28:09.149 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:28:09.149 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:28:09.149 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:09.149 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:09.149 [2024-11-20 13:48:16.179930] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:28:09.149 [2024-11-20 13:48:16.180004] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:09.408 BaseBdev2 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:09.408 [ 00:28:09.408 { 00:28:09.408 "name": "BaseBdev2", 00:28:09.408 "aliases": [ 00:28:09.408 "14c4d398-1394-4fd0-82c7-60394503f9f9" 00:28:09.408 ], 00:28:09.408 "product_name": "Malloc disk", 00:28:09.408 "block_size": 512, 00:28:09.408 "num_blocks": 65536, 00:28:09.408 "uuid": "14c4d398-1394-4fd0-82c7-60394503f9f9", 00:28:09.408 "assigned_rate_limits": { 00:28:09.408 "rw_ios_per_sec": 0, 00:28:09.408 "rw_mbytes_per_sec": 0, 00:28:09.408 "r_mbytes_per_sec": 0, 00:28:09.408 "w_mbytes_per_sec": 0 00:28:09.408 }, 00:28:09.408 "claimed": false, 00:28:09.408 "zoned": false, 00:28:09.408 "supported_io_types": { 00:28:09.408 "read": true, 00:28:09.408 "write": true, 00:28:09.408 "unmap": true, 00:28:09.408 "flush": true, 00:28:09.408 "reset": true, 00:28:09.408 "nvme_admin": false, 00:28:09.408 "nvme_io": false, 00:28:09.408 "nvme_io_md": false, 00:28:09.408 "write_zeroes": true, 00:28:09.408 "zcopy": true, 00:28:09.408 "get_zone_info": false, 00:28:09.408 "zone_management": false, 00:28:09.408 "zone_append": false, 00:28:09.408 "compare": false, 00:28:09.408 "compare_and_write": false, 00:28:09.408 "abort": true, 00:28:09.408 "seek_hole": false, 00:28:09.408 "seek_data": false, 00:28:09.408 "copy": true, 00:28:09.408 "nvme_iov_md": false 00:28:09.408 }, 00:28:09.408 "memory_domains": [ 00:28:09.408 { 00:28:09.408 "dma_device_id": "system", 00:28:09.408 "dma_device_type": 1 00:28:09.408 }, 00:28:09.408 { 00:28:09.408 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:09.408 "dma_device_type": 2 00:28:09.408 } 00:28:09.408 ], 00:28:09.408 "driver_specific": {} 00:28:09.408 } 00:28:09.408 ] 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:09.408 BaseBdev3 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:09.408 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:09.408 [ 00:28:09.408 { 00:28:09.408 "name": "BaseBdev3", 00:28:09.408 "aliases": [ 00:28:09.408 "bd6542e7-93f8-4ecc-afcf-7d0e18d11927" 00:28:09.408 ], 00:28:09.408 "product_name": "Malloc disk", 00:28:09.408 "block_size": 512, 00:28:09.408 "num_blocks": 65536, 00:28:09.408 "uuid": "bd6542e7-93f8-4ecc-afcf-7d0e18d11927", 00:28:09.408 "assigned_rate_limits": { 00:28:09.408 "rw_ios_per_sec": 0, 00:28:09.408 "rw_mbytes_per_sec": 0, 00:28:09.408 "r_mbytes_per_sec": 0, 00:28:09.408 "w_mbytes_per_sec": 0 00:28:09.408 }, 00:28:09.408 "claimed": false, 00:28:09.408 "zoned": false, 00:28:09.408 "supported_io_types": { 00:28:09.408 "read": true, 00:28:09.408 "write": true, 00:28:09.408 "unmap": true, 00:28:09.408 "flush": true, 00:28:09.408 "reset": true, 00:28:09.408 "nvme_admin": false, 00:28:09.408 "nvme_io": false, 00:28:09.408 "nvme_io_md": false, 00:28:09.408 "write_zeroes": true, 00:28:09.408 "zcopy": true, 00:28:09.408 "get_zone_info": false, 00:28:09.408 "zone_management": false, 00:28:09.408 "zone_append": false, 00:28:09.408 "compare": false, 00:28:09.408 "compare_and_write": false, 00:28:09.408 "abort": true, 00:28:09.408 "seek_hole": false, 00:28:09.408 "seek_data": false, 00:28:09.408 "copy": true, 00:28:09.408 "nvme_iov_md": false 00:28:09.408 }, 00:28:09.408 "memory_domains": [ 00:28:09.408 { 00:28:09.408 "dma_device_id": "system", 00:28:09.408 "dma_device_type": 1 00:28:09.408 }, 00:28:09.408 { 00:28:09.408 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:09.408 "dma_device_type": 2 00:28:09.408 } 00:28:09.409 ], 00:28:09.409 "driver_specific": {} 00:28:09.409 } 00:28:09.409 ] 00:28:09.409 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:09.409 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:28:09.409 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:28:09.409 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:28:09.409 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:28:09.409 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:09.409 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:09.409 [2024-11-20 13:48:16.367003] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:28:09.409 [2024-11-20 13:48:16.367055] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:28:09.409 [2024-11-20 13:48:16.367075] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:28:09.409 [2024-11-20 13:48:16.368777] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:28:09.409 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:09.409 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:28:09.409 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:09.409 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:09.409 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:09.409 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:09.409 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:09.409 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:09.409 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:09.409 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:09.409 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:09.409 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:09.409 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:09.409 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:09.409 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:09.409 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:09.409 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:09.409 "name": "Existed_Raid", 00:28:09.409 "uuid": "deb3ef6a-3723-4c90-ae15-4bb912ca5db7", 00:28:09.409 "strip_size_kb": 64, 00:28:09.409 "state": "configuring", 00:28:09.409 "raid_level": "concat", 00:28:09.409 "superblock": true, 00:28:09.409 "num_base_bdevs": 3, 00:28:09.409 "num_base_bdevs_discovered": 2, 00:28:09.409 "num_base_bdevs_operational": 3, 00:28:09.409 "base_bdevs_list": [ 00:28:09.409 { 00:28:09.409 "name": "BaseBdev1", 00:28:09.409 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:09.409 "is_configured": false, 00:28:09.409 "data_offset": 0, 00:28:09.409 "data_size": 0 00:28:09.409 }, 00:28:09.409 { 00:28:09.409 "name": "BaseBdev2", 00:28:09.409 "uuid": "14c4d398-1394-4fd0-82c7-60394503f9f9", 00:28:09.409 "is_configured": true, 00:28:09.409 "data_offset": 2048, 00:28:09.409 "data_size": 63488 00:28:09.409 }, 00:28:09.409 { 00:28:09.409 "name": "BaseBdev3", 00:28:09.409 "uuid": "bd6542e7-93f8-4ecc-afcf-7d0e18d11927", 00:28:09.409 "is_configured": true, 00:28:09.409 "data_offset": 2048, 00:28:09.409 "data_size": 63488 00:28:09.409 } 00:28:09.409 ] 00:28:09.409 }' 00:28:09.409 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:09.409 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:09.676 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:28:09.676 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:09.676 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:09.676 [2024-11-20 13:48:16.691077] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:28:09.676 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:09.676 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:28:09.676 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:09.676 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:09.676 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:09.676 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:09.676 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:09.676 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:09.676 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:09.676 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:09.676 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:09.676 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:09.676 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:09.676 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:09.676 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:09.676 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:09.676 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:09.676 "name": "Existed_Raid", 00:28:09.676 "uuid": "deb3ef6a-3723-4c90-ae15-4bb912ca5db7", 00:28:09.676 "strip_size_kb": 64, 00:28:09.676 "state": "configuring", 00:28:09.676 "raid_level": "concat", 00:28:09.676 "superblock": true, 00:28:09.676 "num_base_bdevs": 3, 00:28:09.676 "num_base_bdevs_discovered": 1, 00:28:09.676 "num_base_bdevs_operational": 3, 00:28:09.676 "base_bdevs_list": [ 00:28:09.676 { 00:28:09.676 "name": "BaseBdev1", 00:28:09.676 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:09.676 "is_configured": false, 00:28:09.676 "data_offset": 0, 00:28:09.676 "data_size": 0 00:28:09.676 }, 00:28:09.676 { 00:28:09.676 "name": null, 00:28:09.676 "uuid": "14c4d398-1394-4fd0-82c7-60394503f9f9", 00:28:09.676 "is_configured": false, 00:28:09.676 "data_offset": 0, 00:28:09.676 "data_size": 63488 00:28:09.676 }, 00:28:09.676 { 00:28:09.676 "name": "BaseBdev3", 00:28:09.676 "uuid": "bd6542e7-93f8-4ecc-afcf-7d0e18d11927", 00:28:09.676 "is_configured": true, 00:28:09.676 "data_offset": 2048, 00:28:09.676 "data_size": 63488 00:28:09.676 } 00:28:09.676 ] 00:28:09.676 }' 00:28:09.934 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:09.934 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:10.192 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:28:10.192 13:48:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:10.192 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:10.192 13:48:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:10.192 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:10.192 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:28:10.192 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:28:10.192 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:10.192 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:10.192 [2024-11-20 13:48:17.039805] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:10.192 BaseBdev1 00:28:10.192 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:10.192 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:28:10.192 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:28:10.192 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:10.193 [ 00:28:10.193 { 00:28:10.193 "name": "BaseBdev1", 00:28:10.193 "aliases": [ 00:28:10.193 "952258b0-03f0-4ba9-803a-ef91586dfd93" 00:28:10.193 ], 00:28:10.193 "product_name": "Malloc disk", 00:28:10.193 "block_size": 512, 00:28:10.193 "num_blocks": 65536, 00:28:10.193 "uuid": "952258b0-03f0-4ba9-803a-ef91586dfd93", 00:28:10.193 "assigned_rate_limits": { 00:28:10.193 "rw_ios_per_sec": 0, 00:28:10.193 "rw_mbytes_per_sec": 0, 00:28:10.193 "r_mbytes_per_sec": 0, 00:28:10.193 "w_mbytes_per_sec": 0 00:28:10.193 }, 00:28:10.193 "claimed": true, 00:28:10.193 "claim_type": "exclusive_write", 00:28:10.193 "zoned": false, 00:28:10.193 "supported_io_types": { 00:28:10.193 "read": true, 00:28:10.193 "write": true, 00:28:10.193 "unmap": true, 00:28:10.193 "flush": true, 00:28:10.193 "reset": true, 00:28:10.193 "nvme_admin": false, 00:28:10.193 "nvme_io": false, 00:28:10.193 "nvme_io_md": false, 00:28:10.193 "write_zeroes": true, 00:28:10.193 "zcopy": true, 00:28:10.193 "get_zone_info": false, 00:28:10.193 "zone_management": false, 00:28:10.193 "zone_append": false, 00:28:10.193 "compare": false, 00:28:10.193 "compare_and_write": false, 00:28:10.193 "abort": true, 00:28:10.193 "seek_hole": false, 00:28:10.193 "seek_data": false, 00:28:10.193 "copy": true, 00:28:10.193 "nvme_iov_md": false 00:28:10.193 }, 00:28:10.193 "memory_domains": [ 00:28:10.193 { 00:28:10.193 "dma_device_id": "system", 00:28:10.193 "dma_device_type": 1 00:28:10.193 }, 00:28:10.193 { 00:28:10.193 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:10.193 "dma_device_type": 2 00:28:10.193 } 00:28:10.193 ], 00:28:10.193 "driver_specific": {} 00:28:10.193 } 00:28:10.193 ] 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:10.193 "name": "Existed_Raid", 00:28:10.193 "uuid": "deb3ef6a-3723-4c90-ae15-4bb912ca5db7", 00:28:10.193 "strip_size_kb": 64, 00:28:10.193 "state": "configuring", 00:28:10.193 "raid_level": "concat", 00:28:10.193 "superblock": true, 00:28:10.193 "num_base_bdevs": 3, 00:28:10.193 "num_base_bdevs_discovered": 2, 00:28:10.193 "num_base_bdevs_operational": 3, 00:28:10.193 "base_bdevs_list": [ 00:28:10.193 { 00:28:10.193 "name": "BaseBdev1", 00:28:10.193 "uuid": "952258b0-03f0-4ba9-803a-ef91586dfd93", 00:28:10.193 "is_configured": true, 00:28:10.193 "data_offset": 2048, 00:28:10.193 "data_size": 63488 00:28:10.193 }, 00:28:10.193 { 00:28:10.193 "name": null, 00:28:10.193 "uuid": "14c4d398-1394-4fd0-82c7-60394503f9f9", 00:28:10.193 "is_configured": false, 00:28:10.193 "data_offset": 0, 00:28:10.193 "data_size": 63488 00:28:10.193 }, 00:28:10.193 { 00:28:10.193 "name": "BaseBdev3", 00:28:10.193 "uuid": "bd6542e7-93f8-4ecc-afcf-7d0e18d11927", 00:28:10.193 "is_configured": true, 00:28:10.193 "data_offset": 2048, 00:28:10.193 "data_size": 63488 00:28:10.193 } 00:28:10.193 ] 00:28:10.193 }' 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:10.193 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:10.453 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:28:10.453 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:10.453 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:10.453 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:10.453 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:10.453 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:28:10.453 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:28:10.453 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:10.453 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:10.453 [2024-11-20 13:48:17.399950] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:28:10.453 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:10.453 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:28:10.453 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:10.453 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:10.453 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:10.453 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:10.453 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:10.453 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:10.453 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:10.453 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:10.453 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:10.453 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:10.453 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:10.453 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:10.453 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:10.453 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:10.453 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:10.453 "name": "Existed_Raid", 00:28:10.453 "uuid": "deb3ef6a-3723-4c90-ae15-4bb912ca5db7", 00:28:10.453 "strip_size_kb": 64, 00:28:10.453 "state": "configuring", 00:28:10.454 "raid_level": "concat", 00:28:10.454 "superblock": true, 00:28:10.454 "num_base_bdevs": 3, 00:28:10.454 "num_base_bdevs_discovered": 1, 00:28:10.454 "num_base_bdevs_operational": 3, 00:28:10.454 "base_bdevs_list": [ 00:28:10.454 { 00:28:10.454 "name": "BaseBdev1", 00:28:10.454 "uuid": "952258b0-03f0-4ba9-803a-ef91586dfd93", 00:28:10.454 "is_configured": true, 00:28:10.454 "data_offset": 2048, 00:28:10.454 "data_size": 63488 00:28:10.454 }, 00:28:10.454 { 00:28:10.454 "name": null, 00:28:10.454 "uuid": "14c4d398-1394-4fd0-82c7-60394503f9f9", 00:28:10.454 "is_configured": false, 00:28:10.454 "data_offset": 0, 00:28:10.454 "data_size": 63488 00:28:10.454 }, 00:28:10.454 { 00:28:10.454 "name": null, 00:28:10.454 "uuid": "bd6542e7-93f8-4ecc-afcf-7d0e18d11927", 00:28:10.454 "is_configured": false, 00:28:10.454 "data_offset": 0, 00:28:10.454 "data_size": 63488 00:28:10.454 } 00:28:10.454 ] 00:28:10.454 }' 00:28:10.454 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:10.454 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:10.711 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:10.711 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:10.711 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:10.711 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:28:10.711 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:10.969 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:28:10.970 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:28:10.970 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:10.970 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:10.970 [2024-11-20 13:48:17.788090] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:28:10.970 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:10.970 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:28:10.970 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:10.970 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:10.970 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:10.970 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:10.970 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:10.970 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:10.970 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:10.970 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:10.970 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:10.970 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:10.970 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:10.970 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:10.970 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:10.970 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:10.970 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:10.970 "name": "Existed_Raid", 00:28:10.970 "uuid": "deb3ef6a-3723-4c90-ae15-4bb912ca5db7", 00:28:10.970 "strip_size_kb": 64, 00:28:10.970 "state": "configuring", 00:28:10.970 "raid_level": "concat", 00:28:10.970 "superblock": true, 00:28:10.970 "num_base_bdevs": 3, 00:28:10.970 "num_base_bdevs_discovered": 2, 00:28:10.970 "num_base_bdevs_operational": 3, 00:28:10.970 "base_bdevs_list": [ 00:28:10.970 { 00:28:10.970 "name": "BaseBdev1", 00:28:10.970 "uuid": "952258b0-03f0-4ba9-803a-ef91586dfd93", 00:28:10.970 "is_configured": true, 00:28:10.970 "data_offset": 2048, 00:28:10.970 "data_size": 63488 00:28:10.970 }, 00:28:10.970 { 00:28:10.970 "name": null, 00:28:10.970 "uuid": "14c4d398-1394-4fd0-82c7-60394503f9f9", 00:28:10.970 "is_configured": false, 00:28:10.970 "data_offset": 0, 00:28:10.970 "data_size": 63488 00:28:10.970 }, 00:28:10.970 { 00:28:10.970 "name": "BaseBdev3", 00:28:10.970 "uuid": "bd6542e7-93f8-4ecc-afcf-7d0e18d11927", 00:28:10.970 "is_configured": true, 00:28:10.970 "data_offset": 2048, 00:28:10.970 "data_size": 63488 00:28:10.970 } 00:28:10.970 ] 00:28:10.970 }' 00:28:10.970 13:48:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:10.970 13:48:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:11.228 [2024-11-20 13:48:18.168145] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:11.228 "name": "Existed_Raid", 00:28:11.228 "uuid": "deb3ef6a-3723-4c90-ae15-4bb912ca5db7", 00:28:11.228 "strip_size_kb": 64, 00:28:11.228 "state": "configuring", 00:28:11.228 "raid_level": "concat", 00:28:11.228 "superblock": true, 00:28:11.228 "num_base_bdevs": 3, 00:28:11.228 "num_base_bdevs_discovered": 1, 00:28:11.228 "num_base_bdevs_operational": 3, 00:28:11.228 "base_bdevs_list": [ 00:28:11.228 { 00:28:11.228 "name": null, 00:28:11.228 "uuid": "952258b0-03f0-4ba9-803a-ef91586dfd93", 00:28:11.228 "is_configured": false, 00:28:11.228 "data_offset": 0, 00:28:11.228 "data_size": 63488 00:28:11.228 }, 00:28:11.228 { 00:28:11.228 "name": null, 00:28:11.228 "uuid": "14c4d398-1394-4fd0-82c7-60394503f9f9", 00:28:11.228 "is_configured": false, 00:28:11.228 "data_offset": 0, 00:28:11.228 "data_size": 63488 00:28:11.228 }, 00:28:11.228 { 00:28:11.228 "name": "BaseBdev3", 00:28:11.228 "uuid": "bd6542e7-93f8-4ecc-afcf-7d0e18d11927", 00:28:11.228 "is_configured": true, 00:28:11.228 "data_offset": 2048, 00:28:11.228 "data_size": 63488 00:28:11.228 } 00:28:11.228 ] 00:28:11.228 }' 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:11.228 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:11.488 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:11.488 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:11.488 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:11.488 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:28:11.745 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:11.745 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:28:11.745 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:28:11.745 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:11.745 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:11.745 [2024-11-20 13:48:18.566198] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:28:11.745 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:11.745 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:28:11.745 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:11.745 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:11.745 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:11.745 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:11.745 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:11.745 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:11.745 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:11.745 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:11.745 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:11.745 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:11.746 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:11.746 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:11.746 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:11.746 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:11.746 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:11.746 "name": "Existed_Raid", 00:28:11.746 "uuid": "deb3ef6a-3723-4c90-ae15-4bb912ca5db7", 00:28:11.746 "strip_size_kb": 64, 00:28:11.746 "state": "configuring", 00:28:11.746 "raid_level": "concat", 00:28:11.746 "superblock": true, 00:28:11.746 "num_base_bdevs": 3, 00:28:11.746 "num_base_bdevs_discovered": 2, 00:28:11.746 "num_base_bdevs_operational": 3, 00:28:11.746 "base_bdevs_list": [ 00:28:11.746 { 00:28:11.746 "name": null, 00:28:11.746 "uuid": "952258b0-03f0-4ba9-803a-ef91586dfd93", 00:28:11.746 "is_configured": false, 00:28:11.746 "data_offset": 0, 00:28:11.746 "data_size": 63488 00:28:11.746 }, 00:28:11.746 { 00:28:11.746 "name": "BaseBdev2", 00:28:11.746 "uuid": "14c4d398-1394-4fd0-82c7-60394503f9f9", 00:28:11.746 "is_configured": true, 00:28:11.746 "data_offset": 2048, 00:28:11.746 "data_size": 63488 00:28:11.746 }, 00:28:11.746 { 00:28:11.746 "name": "BaseBdev3", 00:28:11.746 "uuid": "bd6542e7-93f8-4ecc-afcf-7d0e18d11927", 00:28:11.746 "is_configured": true, 00:28:11.746 "data_offset": 2048, 00:28:11.746 "data_size": 63488 00:28:11.746 } 00:28:11.746 ] 00:28:11.746 }' 00:28:11.746 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:11.746 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:12.004 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:28:12.004 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:12.004 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:12.004 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:12.004 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:12.004 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:28:12.004 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:12.004 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:28:12.004 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:12.004 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:12.004 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:12.004 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 952258b0-03f0-4ba9-803a-ef91586dfd93 00:28:12.004 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:12.004 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:12.004 [2024-11-20 13:48:18.963186] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:28:12.004 [2024-11-20 13:48:18.963398] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:28:12.004 [2024-11-20 13:48:18.963446] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:28:12.004 [2024-11-20 13:48:18.963666] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:28:12.004 [2024-11-20 13:48:18.963790] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:28:12.005 [2024-11-20 13:48:18.963803] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:28:12.005 NewBaseBdev 00:28:12.005 [2024-11-20 13:48:18.963918] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:12.005 [ 00:28:12.005 { 00:28:12.005 "name": "NewBaseBdev", 00:28:12.005 "aliases": [ 00:28:12.005 "952258b0-03f0-4ba9-803a-ef91586dfd93" 00:28:12.005 ], 00:28:12.005 "product_name": "Malloc disk", 00:28:12.005 "block_size": 512, 00:28:12.005 "num_blocks": 65536, 00:28:12.005 "uuid": "952258b0-03f0-4ba9-803a-ef91586dfd93", 00:28:12.005 "assigned_rate_limits": { 00:28:12.005 "rw_ios_per_sec": 0, 00:28:12.005 "rw_mbytes_per_sec": 0, 00:28:12.005 "r_mbytes_per_sec": 0, 00:28:12.005 "w_mbytes_per_sec": 0 00:28:12.005 }, 00:28:12.005 "claimed": true, 00:28:12.005 "claim_type": "exclusive_write", 00:28:12.005 "zoned": false, 00:28:12.005 "supported_io_types": { 00:28:12.005 "read": true, 00:28:12.005 "write": true, 00:28:12.005 "unmap": true, 00:28:12.005 "flush": true, 00:28:12.005 "reset": true, 00:28:12.005 "nvme_admin": false, 00:28:12.005 "nvme_io": false, 00:28:12.005 "nvme_io_md": false, 00:28:12.005 "write_zeroes": true, 00:28:12.005 "zcopy": true, 00:28:12.005 "get_zone_info": false, 00:28:12.005 "zone_management": false, 00:28:12.005 "zone_append": false, 00:28:12.005 "compare": false, 00:28:12.005 "compare_and_write": false, 00:28:12.005 "abort": true, 00:28:12.005 "seek_hole": false, 00:28:12.005 "seek_data": false, 00:28:12.005 "copy": true, 00:28:12.005 "nvme_iov_md": false 00:28:12.005 }, 00:28:12.005 "memory_domains": [ 00:28:12.005 { 00:28:12.005 "dma_device_id": "system", 00:28:12.005 "dma_device_type": 1 00:28:12.005 }, 00:28:12.005 { 00:28:12.005 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:12.005 "dma_device_type": 2 00:28:12.005 } 00:28:12.005 ], 00:28:12.005 "driver_specific": {} 00:28:12.005 } 00:28:12.005 ] 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:12.005 13:48:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:12.005 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:12.005 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:12.005 "name": "Existed_Raid", 00:28:12.005 "uuid": "deb3ef6a-3723-4c90-ae15-4bb912ca5db7", 00:28:12.005 "strip_size_kb": 64, 00:28:12.005 "state": "online", 00:28:12.005 "raid_level": "concat", 00:28:12.005 "superblock": true, 00:28:12.005 "num_base_bdevs": 3, 00:28:12.005 "num_base_bdevs_discovered": 3, 00:28:12.005 "num_base_bdevs_operational": 3, 00:28:12.005 "base_bdevs_list": [ 00:28:12.005 { 00:28:12.005 "name": "NewBaseBdev", 00:28:12.005 "uuid": "952258b0-03f0-4ba9-803a-ef91586dfd93", 00:28:12.005 "is_configured": true, 00:28:12.005 "data_offset": 2048, 00:28:12.005 "data_size": 63488 00:28:12.005 }, 00:28:12.005 { 00:28:12.005 "name": "BaseBdev2", 00:28:12.005 "uuid": "14c4d398-1394-4fd0-82c7-60394503f9f9", 00:28:12.005 "is_configured": true, 00:28:12.005 "data_offset": 2048, 00:28:12.005 "data_size": 63488 00:28:12.005 }, 00:28:12.005 { 00:28:12.005 "name": "BaseBdev3", 00:28:12.005 "uuid": "bd6542e7-93f8-4ecc-afcf-7d0e18d11927", 00:28:12.005 "is_configured": true, 00:28:12.005 "data_offset": 2048, 00:28:12.005 "data_size": 63488 00:28:12.005 } 00:28:12.005 ] 00:28:12.005 }' 00:28:12.005 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:12.005 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:12.264 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:28:12.264 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:28:12.264 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:28:12.264 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:28:12.264 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:28:12.264 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:28:12.264 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:28:12.264 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:28:12.264 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:12.264 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:12.264 [2024-11-20 13:48:19.287610] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:12.264 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:12.264 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:12.264 "name": "Existed_Raid", 00:28:12.264 "aliases": [ 00:28:12.264 "deb3ef6a-3723-4c90-ae15-4bb912ca5db7" 00:28:12.264 ], 00:28:12.264 "product_name": "Raid Volume", 00:28:12.264 "block_size": 512, 00:28:12.264 "num_blocks": 190464, 00:28:12.264 "uuid": "deb3ef6a-3723-4c90-ae15-4bb912ca5db7", 00:28:12.264 "assigned_rate_limits": { 00:28:12.264 "rw_ios_per_sec": 0, 00:28:12.264 "rw_mbytes_per_sec": 0, 00:28:12.264 "r_mbytes_per_sec": 0, 00:28:12.264 "w_mbytes_per_sec": 0 00:28:12.264 }, 00:28:12.264 "claimed": false, 00:28:12.264 "zoned": false, 00:28:12.264 "supported_io_types": { 00:28:12.264 "read": true, 00:28:12.264 "write": true, 00:28:12.264 "unmap": true, 00:28:12.264 "flush": true, 00:28:12.264 "reset": true, 00:28:12.264 "nvme_admin": false, 00:28:12.264 "nvme_io": false, 00:28:12.264 "nvme_io_md": false, 00:28:12.264 "write_zeroes": true, 00:28:12.264 "zcopy": false, 00:28:12.264 "get_zone_info": false, 00:28:12.264 "zone_management": false, 00:28:12.264 "zone_append": false, 00:28:12.264 "compare": false, 00:28:12.264 "compare_and_write": false, 00:28:12.264 "abort": false, 00:28:12.264 "seek_hole": false, 00:28:12.264 "seek_data": false, 00:28:12.264 "copy": false, 00:28:12.264 "nvme_iov_md": false 00:28:12.264 }, 00:28:12.264 "memory_domains": [ 00:28:12.264 { 00:28:12.264 "dma_device_id": "system", 00:28:12.264 "dma_device_type": 1 00:28:12.264 }, 00:28:12.264 { 00:28:12.264 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:12.264 "dma_device_type": 2 00:28:12.264 }, 00:28:12.264 { 00:28:12.264 "dma_device_id": "system", 00:28:12.264 "dma_device_type": 1 00:28:12.264 }, 00:28:12.264 { 00:28:12.264 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:12.264 "dma_device_type": 2 00:28:12.264 }, 00:28:12.264 { 00:28:12.264 "dma_device_id": "system", 00:28:12.264 "dma_device_type": 1 00:28:12.264 }, 00:28:12.264 { 00:28:12.264 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:12.264 "dma_device_type": 2 00:28:12.264 } 00:28:12.264 ], 00:28:12.264 "driver_specific": { 00:28:12.264 "raid": { 00:28:12.264 "uuid": "deb3ef6a-3723-4c90-ae15-4bb912ca5db7", 00:28:12.264 "strip_size_kb": 64, 00:28:12.264 "state": "online", 00:28:12.264 "raid_level": "concat", 00:28:12.264 "superblock": true, 00:28:12.264 "num_base_bdevs": 3, 00:28:12.264 "num_base_bdevs_discovered": 3, 00:28:12.264 "num_base_bdevs_operational": 3, 00:28:12.264 "base_bdevs_list": [ 00:28:12.264 { 00:28:12.264 "name": "NewBaseBdev", 00:28:12.264 "uuid": "952258b0-03f0-4ba9-803a-ef91586dfd93", 00:28:12.264 "is_configured": true, 00:28:12.264 "data_offset": 2048, 00:28:12.264 "data_size": 63488 00:28:12.264 }, 00:28:12.264 { 00:28:12.264 "name": "BaseBdev2", 00:28:12.264 "uuid": "14c4d398-1394-4fd0-82c7-60394503f9f9", 00:28:12.264 "is_configured": true, 00:28:12.264 "data_offset": 2048, 00:28:12.264 "data_size": 63488 00:28:12.264 }, 00:28:12.264 { 00:28:12.264 "name": "BaseBdev3", 00:28:12.264 "uuid": "bd6542e7-93f8-4ecc-afcf-7d0e18d11927", 00:28:12.264 "is_configured": true, 00:28:12.264 "data_offset": 2048, 00:28:12.264 "data_size": 63488 00:28:12.264 } 00:28:12.264 ] 00:28:12.264 } 00:28:12.264 } 00:28:12.264 }' 00:28:12.264 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:28:12.564 BaseBdev2 00:28:12.564 BaseBdev3' 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:12.564 [2024-11-20 13:48:19.487361] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:28:12.564 [2024-11-20 13:48:19.487395] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:28:12.564 [2024-11-20 13:48:19.487485] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:12.564 [2024-11-20 13:48:19.487548] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:12.564 [2024-11-20 13:48:19.487563] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 64636 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 64636 ']' 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 64636 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 64636 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:28:12.564 killing process with pid 64636 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 64636' 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 64636 00:28:12.564 [2024-11-20 13:48:19.521339] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:28:12.564 13:48:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 64636 00:28:12.820 [2024-11-20 13:48:19.681424] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:28:13.409 13:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:28:13.409 00:28:13.409 real 0m7.535s 00:28:13.409 user 0m12.028s 00:28:13.409 sys 0m1.346s 00:28:13.409 13:48:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:28:13.409 13:48:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:13.409 ************************************ 00:28:13.409 END TEST raid_state_function_test_sb 00:28:13.409 ************************************ 00:28:13.409 13:48:20 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 3 00:28:13.409 13:48:20 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:28:13.409 13:48:20 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:28:13.409 13:48:20 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:28:13.409 ************************************ 00:28:13.409 START TEST raid_superblock_test 00:28:13.409 ************************************ 00:28:13.409 13:48:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test concat 3 00:28:13.409 13:48:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:28:13.409 13:48:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:28:13.409 13:48:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:28:13.409 13:48:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:28:13.409 13:48:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:28:13.409 13:48:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:28:13.409 13:48:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:28:13.409 13:48:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:28:13.409 13:48:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:28:13.409 13:48:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:28:13.409 13:48:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:28:13.409 13:48:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:28:13.409 13:48:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:28:13.409 13:48:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:28:13.409 13:48:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:28:13.409 13:48:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:28:13.409 13:48:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=65223 00:28:13.409 13:48:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 65223 00:28:13.409 13:48:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 65223 ']' 00:28:13.409 13:48:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:28:13.409 13:48:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:28:13.409 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:28:13.409 13:48:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:28:13.409 13:48:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:28:13.409 13:48:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:28:13.409 13:48:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:13.409 [2024-11-20 13:48:20.442719] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:28:13.409 [2024-11-20 13:48:20.442902] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid65223 ] 00:28:13.668 [2024-11-20 13:48:20.609561] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:13.668 [2024-11-20 13:48:20.717859] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:28:13.926 [2024-11-20 13:48:20.843142] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:13.926 [2024-11-20 13:48:20.843217] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:14.492 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:28:14.492 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:28:14.492 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:28:14.492 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:28:14.492 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:28:14.492 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:28:14.492 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:28:14.492 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:28:14.492 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:28:14.492 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:28:14.492 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:14.493 malloc1 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:14.493 [2024-11-20 13:48:21.322828] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:28:14.493 [2024-11-20 13:48:21.322918] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:14.493 [2024-11-20 13:48:21.322941] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:28:14.493 [2024-11-20 13:48:21.322950] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:14.493 [2024-11-20 13:48:21.325098] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:14.493 [2024-11-20 13:48:21.325151] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:28:14.493 pt1 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:14.493 malloc2 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:14.493 [2024-11-20 13:48:21.357853] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:28:14.493 [2024-11-20 13:48:21.357930] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:14.493 [2024-11-20 13:48:21.357957] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:28:14.493 [2024-11-20 13:48:21.357966] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:14.493 [2024-11-20 13:48:21.360057] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:14.493 [2024-11-20 13:48:21.360106] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:28:14.493 pt2 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:14.493 malloc3 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:14.493 [2024-11-20 13:48:21.408491] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:28:14.493 [2024-11-20 13:48:21.408566] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:14.493 [2024-11-20 13:48:21.408592] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:28:14.493 [2024-11-20 13:48:21.408603] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:14.493 [2024-11-20 13:48:21.410652] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:14.493 [2024-11-20 13:48:21.410696] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:28:14.493 pt3 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:14.493 [2024-11-20 13:48:21.416552] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:28:14.493 [2024-11-20 13:48:21.418313] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:28:14.493 [2024-11-20 13:48:21.418383] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:28:14.493 [2024-11-20 13:48:21.418538] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:28:14.493 [2024-11-20 13:48:21.418556] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:28:14.493 [2024-11-20 13:48:21.418818] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:28:14.493 [2024-11-20 13:48:21.418991] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:28:14.493 [2024-11-20 13:48:21.419005] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:28:14.493 [2024-11-20 13:48:21.419162] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:14.493 "name": "raid_bdev1", 00:28:14.493 "uuid": "2d457e76-58af-499f-8300-09be0bf8e78a", 00:28:14.493 "strip_size_kb": 64, 00:28:14.493 "state": "online", 00:28:14.493 "raid_level": "concat", 00:28:14.493 "superblock": true, 00:28:14.493 "num_base_bdevs": 3, 00:28:14.493 "num_base_bdevs_discovered": 3, 00:28:14.493 "num_base_bdevs_operational": 3, 00:28:14.493 "base_bdevs_list": [ 00:28:14.493 { 00:28:14.493 "name": "pt1", 00:28:14.493 "uuid": "00000000-0000-0000-0000-000000000001", 00:28:14.493 "is_configured": true, 00:28:14.493 "data_offset": 2048, 00:28:14.493 "data_size": 63488 00:28:14.493 }, 00:28:14.493 { 00:28:14.493 "name": "pt2", 00:28:14.493 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:14.493 "is_configured": true, 00:28:14.493 "data_offset": 2048, 00:28:14.493 "data_size": 63488 00:28:14.493 }, 00:28:14.493 { 00:28:14.493 "name": "pt3", 00:28:14.493 "uuid": "00000000-0000-0000-0000-000000000003", 00:28:14.493 "is_configured": true, 00:28:14.493 "data_offset": 2048, 00:28:14.493 "data_size": 63488 00:28:14.493 } 00:28:14.493 ] 00:28:14.493 }' 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:14.493 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:14.752 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:28:14.752 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:28:14.752 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:28:14.752 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:28:14.752 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:28:14.752 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:28:14.752 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:28:14.752 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:14.752 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:14.752 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:28:14.752 [2024-11-20 13:48:21.728860] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:14.752 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:14.752 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:14.752 "name": "raid_bdev1", 00:28:14.752 "aliases": [ 00:28:14.752 "2d457e76-58af-499f-8300-09be0bf8e78a" 00:28:14.752 ], 00:28:14.752 "product_name": "Raid Volume", 00:28:14.752 "block_size": 512, 00:28:14.752 "num_blocks": 190464, 00:28:14.752 "uuid": "2d457e76-58af-499f-8300-09be0bf8e78a", 00:28:14.752 "assigned_rate_limits": { 00:28:14.752 "rw_ios_per_sec": 0, 00:28:14.752 "rw_mbytes_per_sec": 0, 00:28:14.752 "r_mbytes_per_sec": 0, 00:28:14.752 "w_mbytes_per_sec": 0 00:28:14.752 }, 00:28:14.752 "claimed": false, 00:28:14.752 "zoned": false, 00:28:14.752 "supported_io_types": { 00:28:14.752 "read": true, 00:28:14.752 "write": true, 00:28:14.752 "unmap": true, 00:28:14.752 "flush": true, 00:28:14.752 "reset": true, 00:28:14.752 "nvme_admin": false, 00:28:14.752 "nvme_io": false, 00:28:14.752 "nvme_io_md": false, 00:28:14.752 "write_zeroes": true, 00:28:14.752 "zcopy": false, 00:28:14.752 "get_zone_info": false, 00:28:14.752 "zone_management": false, 00:28:14.752 "zone_append": false, 00:28:14.752 "compare": false, 00:28:14.752 "compare_and_write": false, 00:28:14.752 "abort": false, 00:28:14.752 "seek_hole": false, 00:28:14.752 "seek_data": false, 00:28:14.752 "copy": false, 00:28:14.752 "nvme_iov_md": false 00:28:14.752 }, 00:28:14.752 "memory_domains": [ 00:28:14.752 { 00:28:14.752 "dma_device_id": "system", 00:28:14.752 "dma_device_type": 1 00:28:14.752 }, 00:28:14.752 { 00:28:14.752 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:14.752 "dma_device_type": 2 00:28:14.752 }, 00:28:14.752 { 00:28:14.752 "dma_device_id": "system", 00:28:14.752 "dma_device_type": 1 00:28:14.752 }, 00:28:14.752 { 00:28:14.752 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:14.752 "dma_device_type": 2 00:28:14.752 }, 00:28:14.752 { 00:28:14.752 "dma_device_id": "system", 00:28:14.752 "dma_device_type": 1 00:28:14.752 }, 00:28:14.752 { 00:28:14.752 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:14.752 "dma_device_type": 2 00:28:14.752 } 00:28:14.752 ], 00:28:14.752 "driver_specific": { 00:28:14.752 "raid": { 00:28:14.752 "uuid": "2d457e76-58af-499f-8300-09be0bf8e78a", 00:28:14.752 "strip_size_kb": 64, 00:28:14.752 "state": "online", 00:28:14.752 "raid_level": "concat", 00:28:14.752 "superblock": true, 00:28:14.752 "num_base_bdevs": 3, 00:28:14.752 "num_base_bdevs_discovered": 3, 00:28:14.752 "num_base_bdevs_operational": 3, 00:28:14.752 "base_bdevs_list": [ 00:28:14.752 { 00:28:14.752 "name": "pt1", 00:28:14.752 "uuid": "00000000-0000-0000-0000-000000000001", 00:28:14.752 "is_configured": true, 00:28:14.752 "data_offset": 2048, 00:28:14.752 "data_size": 63488 00:28:14.752 }, 00:28:14.752 { 00:28:14.752 "name": "pt2", 00:28:14.752 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:14.752 "is_configured": true, 00:28:14.752 "data_offset": 2048, 00:28:14.752 "data_size": 63488 00:28:14.752 }, 00:28:14.752 { 00:28:14.752 "name": "pt3", 00:28:14.752 "uuid": "00000000-0000-0000-0000-000000000003", 00:28:14.752 "is_configured": true, 00:28:14.752 "data_offset": 2048, 00:28:14.752 "data_size": 63488 00:28:14.752 } 00:28:14.752 ] 00:28:14.752 } 00:28:14.752 } 00:28:14.752 }' 00:28:14.752 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:28:14.752 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:28:14.752 pt2 00:28:14.752 pt3' 00:28:14.752 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:15.011 [2024-11-20 13:48:21.936880] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=2d457e76-58af-499f-8300-09be0bf8e78a 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 2d457e76-58af-499f-8300-09be0bf8e78a ']' 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:15.011 [2024-11-20 13:48:21.960613] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:28:15.011 [2024-11-20 13:48:21.960656] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:28:15.011 [2024-11-20 13:48:21.960741] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:15.011 [2024-11-20 13:48:21.960817] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:15.011 [2024-11-20 13:48:21.960832] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:15.011 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:28:15.012 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:28:15.012 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:28:15.012 13:48:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:28:15.012 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:15.012 13:48:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:15.012 [2024-11-20 13:48:22.060679] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:28:15.012 [2024-11-20 13:48:22.062528] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:28:15.012 [2024-11-20 13:48:22.062583] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:28:15.012 [2024-11-20 13:48:22.062635] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:28:15.012 [2024-11-20 13:48:22.062691] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:28:15.012 [2024-11-20 13:48:22.062708] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:28:15.012 [2024-11-20 13:48:22.062723] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:28:15.012 [2024-11-20 13:48:22.062732] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:28:15.012 request: 00:28:15.012 { 00:28:15.012 "name": "raid_bdev1", 00:28:15.012 "raid_level": "concat", 00:28:15.012 "base_bdevs": [ 00:28:15.012 "malloc1", 00:28:15.012 "malloc2", 00:28:15.012 "malloc3" 00:28:15.012 ], 00:28:15.012 "strip_size_kb": 64, 00:28:15.012 "superblock": false, 00:28:15.012 "method": "bdev_raid_create", 00:28:15.012 "req_id": 1 00:28:15.012 } 00:28:15.012 Got JSON-RPC error response 00:28:15.012 response: 00:28:15.012 { 00:28:15.012 "code": -17, 00:28:15.012 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:28:15.012 } 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:28:15.012 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:15.272 [2024-11-20 13:48:22.104637] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:28:15.272 [2024-11-20 13:48:22.104712] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:15.272 [2024-11-20 13:48:22.104732] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:28:15.272 [2024-11-20 13:48:22.104740] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:15.272 [2024-11-20 13:48:22.106829] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:15.272 [2024-11-20 13:48:22.106868] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:28:15.272 [2024-11-20 13:48:22.106956] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:28:15.272 [2024-11-20 13:48:22.107016] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:28:15.272 pt1 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 3 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:15.272 "name": "raid_bdev1", 00:28:15.272 "uuid": "2d457e76-58af-499f-8300-09be0bf8e78a", 00:28:15.272 "strip_size_kb": 64, 00:28:15.272 "state": "configuring", 00:28:15.272 "raid_level": "concat", 00:28:15.272 "superblock": true, 00:28:15.272 "num_base_bdevs": 3, 00:28:15.272 "num_base_bdevs_discovered": 1, 00:28:15.272 "num_base_bdevs_operational": 3, 00:28:15.272 "base_bdevs_list": [ 00:28:15.272 { 00:28:15.272 "name": "pt1", 00:28:15.272 "uuid": "00000000-0000-0000-0000-000000000001", 00:28:15.272 "is_configured": true, 00:28:15.272 "data_offset": 2048, 00:28:15.272 "data_size": 63488 00:28:15.272 }, 00:28:15.272 { 00:28:15.272 "name": null, 00:28:15.272 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:15.272 "is_configured": false, 00:28:15.272 "data_offset": 2048, 00:28:15.272 "data_size": 63488 00:28:15.272 }, 00:28:15.272 { 00:28:15.272 "name": null, 00:28:15.272 "uuid": "00000000-0000-0000-0000-000000000003", 00:28:15.272 "is_configured": false, 00:28:15.272 "data_offset": 2048, 00:28:15.272 "data_size": 63488 00:28:15.272 } 00:28:15.272 ] 00:28:15.272 }' 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:15.272 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:15.540 [2024-11-20 13:48:22.416693] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:28:15.540 [2024-11-20 13:48:22.416776] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:15.540 [2024-11-20 13:48:22.416803] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:28:15.540 [2024-11-20 13:48:22.416812] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:15.540 [2024-11-20 13:48:22.417251] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:15.540 [2024-11-20 13:48:22.417271] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:28:15.540 [2024-11-20 13:48:22.417357] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:28:15.540 [2024-11-20 13:48:22.417381] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:28:15.540 pt2 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:15.540 [2024-11-20 13:48:22.424725] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 3 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:15.540 "name": "raid_bdev1", 00:28:15.540 "uuid": "2d457e76-58af-499f-8300-09be0bf8e78a", 00:28:15.540 "strip_size_kb": 64, 00:28:15.540 "state": "configuring", 00:28:15.540 "raid_level": "concat", 00:28:15.540 "superblock": true, 00:28:15.540 "num_base_bdevs": 3, 00:28:15.540 "num_base_bdevs_discovered": 1, 00:28:15.540 "num_base_bdevs_operational": 3, 00:28:15.540 "base_bdevs_list": [ 00:28:15.540 { 00:28:15.540 "name": "pt1", 00:28:15.540 "uuid": "00000000-0000-0000-0000-000000000001", 00:28:15.540 "is_configured": true, 00:28:15.540 "data_offset": 2048, 00:28:15.540 "data_size": 63488 00:28:15.540 }, 00:28:15.540 { 00:28:15.540 "name": null, 00:28:15.540 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:15.540 "is_configured": false, 00:28:15.540 "data_offset": 0, 00:28:15.540 "data_size": 63488 00:28:15.540 }, 00:28:15.540 { 00:28:15.540 "name": null, 00:28:15.540 "uuid": "00000000-0000-0000-0000-000000000003", 00:28:15.540 "is_configured": false, 00:28:15.540 "data_offset": 2048, 00:28:15.540 "data_size": 63488 00:28:15.540 } 00:28:15.540 ] 00:28:15.540 }' 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:15.540 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:15.798 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:28:15.798 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:28:15.798 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:28:15.798 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:15.798 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:15.798 [2024-11-20 13:48:22.732720] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:28:15.798 [2024-11-20 13:48:22.732799] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:15.798 [2024-11-20 13:48:22.732816] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:28:15.798 [2024-11-20 13:48:22.732825] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:15.798 [2024-11-20 13:48:22.733282] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:15.798 [2024-11-20 13:48:22.733309] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:28:15.798 [2024-11-20 13:48:22.733385] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:28:15.798 [2024-11-20 13:48:22.733406] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:28:15.798 pt2 00:28:15.798 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:15.798 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:28:15.798 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:28:15.798 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:28:15.798 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:15.799 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:15.799 [2024-11-20 13:48:22.740724] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:28:15.799 [2024-11-20 13:48:22.740782] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:15.799 [2024-11-20 13:48:22.740797] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:28:15.799 [2024-11-20 13:48:22.740805] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:15.799 [2024-11-20 13:48:22.741223] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:15.799 [2024-11-20 13:48:22.741249] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:28:15.799 [2024-11-20 13:48:22.741319] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:28:15.799 [2024-11-20 13:48:22.741338] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:28:15.799 [2024-11-20 13:48:22.741451] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:28:15.799 [2024-11-20 13:48:22.741467] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:28:15.799 [2024-11-20 13:48:22.741684] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:28:15.799 [2024-11-20 13:48:22.741821] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:28:15.799 [2024-11-20 13:48:22.741833] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:28:15.799 [2024-11-20 13:48:22.741952] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:15.799 pt3 00:28:15.799 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:15.799 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:28:15.799 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:28:15.799 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:28:15.799 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:15.799 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:15.799 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:15.799 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:15.799 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:15.799 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:15.799 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:15.799 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:15.799 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:15.799 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:15.799 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:15.799 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:15.799 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:15.799 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:15.799 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:15.799 "name": "raid_bdev1", 00:28:15.799 "uuid": "2d457e76-58af-499f-8300-09be0bf8e78a", 00:28:15.799 "strip_size_kb": 64, 00:28:15.799 "state": "online", 00:28:15.799 "raid_level": "concat", 00:28:15.799 "superblock": true, 00:28:15.799 "num_base_bdevs": 3, 00:28:15.799 "num_base_bdevs_discovered": 3, 00:28:15.799 "num_base_bdevs_operational": 3, 00:28:15.799 "base_bdevs_list": [ 00:28:15.799 { 00:28:15.799 "name": "pt1", 00:28:15.799 "uuid": "00000000-0000-0000-0000-000000000001", 00:28:15.799 "is_configured": true, 00:28:15.799 "data_offset": 2048, 00:28:15.799 "data_size": 63488 00:28:15.799 }, 00:28:15.799 { 00:28:15.799 "name": "pt2", 00:28:15.799 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:15.799 "is_configured": true, 00:28:15.799 "data_offset": 2048, 00:28:15.799 "data_size": 63488 00:28:15.799 }, 00:28:15.799 { 00:28:15.799 "name": "pt3", 00:28:15.799 "uuid": "00000000-0000-0000-0000-000000000003", 00:28:15.799 "is_configured": true, 00:28:15.799 "data_offset": 2048, 00:28:15.799 "data_size": 63488 00:28:15.799 } 00:28:15.799 ] 00:28:15.799 }' 00:28:15.799 13:48:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:15.799 13:48:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:16.057 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:28:16.057 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:28:16.057 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:28:16.057 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:28:16.057 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:28:16.057 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:28:16.057 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:28:16.057 13:48:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:16.057 13:48:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:16.057 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:28:16.057 [2024-11-20 13:48:23.061106] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:16.057 13:48:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:16.057 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:16.057 "name": "raid_bdev1", 00:28:16.057 "aliases": [ 00:28:16.057 "2d457e76-58af-499f-8300-09be0bf8e78a" 00:28:16.057 ], 00:28:16.057 "product_name": "Raid Volume", 00:28:16.057 "block_size": 512, 00:28:16.057 "num_blocks": 190464, 00:28:16.057 "uuid": "2d457e76-58af-499f-8300-09be0bf8e78a", 00:28:16.057 "assigned_rate_limits": { 00:28:16.057 "rw_ios_per_sec": 0, 00:28:16.057 "rw_mbytes_per_sec": 0, 00:28:16.057 "r_mbytes_per_sec": 0, 00:28:16.057 "w_mbytes_per_sec": 0 00:28:16.057 }, 00:28:16.057 "claimed": false, 00:28:16.057 "zoned": false, 00:28:16.057 "supported_io_types": { 00:28:16.057 "read": true, 00:28:16.057 "write": true, 00:28:16.057 "unmap": true, 00:28:16.057 "flush": true, 00:28:16.057 "reset": true, 00:28:16.057 "nvme_admin": false, 00:28:16.057 "nvme_io": false, 00:28:16.057 "nvme_io_md": false, 00:28:16.057 "write_zeroes": true, 00:28:16.057 "zcopy": false, 00:28:16.057 "get_zone_info": false, 00:28:16.057 "zone_management": false, 00:28:16.057 "zone_append": false, 00:28:16.057 "compare": false, 00:28:16.057 "compare_and_write": false, 00:28:16.057 "abort": false, 00:28:16.057 "seek_hole": false, 00:28:16.057 "seek_data": false, 00:28:16.057 "copy": false, 00:28:16.057 "nvme_iov_md": false 00:28:16.057 }, 00:28:16.057 "memory_domains": [ 00:28:16.057 { 00:28:16.057 "dma_device_id": "system", 00:28:16.057 "dma_device_type": 1 00:28:16.057 }, 00:28:16.057 { 00:28:16.057 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:16.057 "dma_device_type": 2 00:28:16.057 }, 00:28:16.057 { 00:28:16.057 "dma_device_id": "system", 00:28:16.057 "dma_device_type": 1 00:28:16.057 }, 00:28:16.057 { 00:28:16.057 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:16.057 "dma_device_type": 2 00:28:16.057 }, 00:28:16.057 { 00:28:16.057 "dma_device_id": "system", 00:28:16.057 "dma_device_type": 1 00:28:16.057 }, 00:28:16.057 { 00:28:16.057 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:16.057 "dma_device_type": 2 00:28:16.057 } 00:28:16.057 ], 00:28:16.057 "driver_specific": { 00:28:16.057 "raid": { 00:28:16.057 "uuid": "2d457e76-58af-499f-8300-09be0bf8e78a", 00:28:16.057 "strip_size_kb": 64, 00:28:16.057 "state": "online", 00:28:16.057 "raid_level": "concat", 00:28:16.057 "superblock": true, 00:28:16.057 "num_base_bdevs": 3, 00:28:16.057 "num_base_bdevs_discovered": 3, 00:28:16.057 "num_base_bdevs_operational": 3, 00:28:16.057 "base_bdevs_list": [ 00:28:16.057 { 00:28:16.057 "name": "pt1", 00:28:16.057 "uuid": "00000000-0000-0000-0000-000000000001", 00:28:16.057 "is_configured": true, 00:28:16.057 "data_offset": 2048, 00:28:16.057 "data_size": 63488 00:28:16.057 }, 00:28:16.057 { 00:28:16.057 "name": "pt2", 00:28:16.057 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:16.057 "is_configured": true, 00:28:16.057 "data_offset": 2048, 00:28:16.057 "data_size": 63488 00:28:16.057 }, 00:28:16.057 { 00:28:16.057 "name": "pt3", 00:28:16.057 "uuid": "00000000-0000-0000-0000-000000000003", 00:28:16.057 "is_configured": true, 00:28:16.057 "data_offset": 2048, 00:28:16.057 "data_size": 63488 00:28:16.057 } 00:28:16.057 ] 00:28:16.057 } 00:28:16.057 } 00:28:16.057 }' 00:28:16.057 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:28:16.316 pt2 00:28:16.316 pt3' 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:28:16.316 [2024-11-20 13:48:23.273117] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 2d457e76-58af-499f-8300-09be0bf8e78a '!=' 2d457e76-58af-499f-8300-09be0bf8e78a ']' 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 65223 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 65223 ']' 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 65223 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 65223 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:28:16.316 killing process with pid 65223 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 65223' 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 65223 00:28:16.316 13:48:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 65223 00:28:16.316 [2024-11-20 13:48:23.324753] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:28:16.316 [2024-11-20 13:48:23.324866] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:16.316 [2024-11-20 13:48:23.324936] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:16.316 [2024-11-20 13:48:23.324947] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:28:16.575 [2024-11-20 13:48:23.487336] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:28:17.141 13:48:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:28:17.141 00:28:17.141 real 0m3.761s 00:28:17.141 user 0m5.373s 00:28:17.141 sys 0m0.663s 00:28:17.141 13:48:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:28:17.141 13:48:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:17.141 ************************************ 00:28:17.141 END TEST raid_superblock_test 00:28:17.141 ************************************ 00:28:17.141 13:48:24 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 3 read 00:28:17.141 13:48:24 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:28:17.141 13:48:24 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:28:17.141 13:48:24 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:28:17.141 ************************************ 00:28:17.141 START TEST raid_read_error_test 00:28:17.141 ************************************ 00:28:17.141 13:48:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 3 read 00:28:17.141 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:28:17.141 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:28:17.141 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:28:17.141 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:28:17.141 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:28:17.141 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:28:17.141 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:28:17.141 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:28:17.141 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:28:17.141 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:28:17.141 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:28:17.141 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:28:17.141 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:28:17.141 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:28:17.141 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:28:17.142 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:28:17.142 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:28:17.142 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:28:17.142 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:28:17.142 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:28:17.142 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:28:17.142 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:28:17.142 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:28:17.142 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:28:17.142 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:28:17.142 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.kYptKtm8NQ 00:28:17.142 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=65464 00:28:17.142 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 65464 00:28:17.142 13:48:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 65464 ']' 00:28:17.142 13:48:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:28:17.142 13:48:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:28:17.142 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:28:17.142 13:48:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:28:17.142 13:48:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:28:17.142 13:48:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:17.142 13:48:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:28:17.399 [2024-11-20 13:48:24.232035] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:28:17.400 [2024-11-20 13:48:24.232155] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid65464 ] 00:28:17.400 [2024-11-20 13:48:24.384587] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:17.657 [2024-11-20 13:48:24.504920] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:28:17.657 [2024-11-20 13:48:24.653768] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:17.657 [2024-11-20 13:48:24.653826] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:18.222 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:28:18.222 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:28:18.222 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:18.223 BaseBdev1_malloc 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:18.223 true 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:18.223 [2024-11-20 13:48:25.153482] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:28:18.223 [2024-11-20 13:48:25.153558] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:18.223 [2024-11-20 13:48:25.153582] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:28:18.223 [2024-11-20 13:48:25.153594] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:18.223 [2024-11-20 13:48:25.155961] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:18.223 [2024-11-20 13:48:25.156033] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:28:18.223 BaseBdev1 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:18.223 BaseBdev2_malloc 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:18.223 true 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:18.223 [2024-11-20 13:48:25.199669] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:28:18.223 [2024-11-20 13:48:25.199742] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:18.223 [2024-11-20 13:48:25.199764] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:28:18.223 [2024-11-20 13:48:25.199776] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:18.223 [2024-11-20 13:48:25.202205] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:18.223 [2024-11-20 13:48:25.202250] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:28:18.223 BaseBdev2 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:18.223 BaseBdev3_malloc 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:18.223 true 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:18.223 [2024-11-20 13:48:25.270156] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:28:18.223 [2024-11-20 13:48:25.270233] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:18.223 [2024-11-20 13:48:25.270255] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:28:18.223 [2024-11-20 13:48:25.270267] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:18.223 [2024-11-20 13:48:25.273730] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:18.223 [2024-11-20 13:48:25.273789] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:28:18.223 BaseBdev3 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:18.223 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:18.223 [2024-11-20 13:48:25.278474] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:18.481 [2024-11-20 13:48:25.280599] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:28:18.481 [2024-11-20 13:48:25.280697] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:28:18.481 [2024-11-20 13:48:25.280923] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:28:18.481 [2024-11-20 13:48:25.280942] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:28:18.481 [2024-11-20 13:48:25.281267] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006560 00:28:18.481 [2024-11-20 13:48:25.281440] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:28:18.481 [2024-11-20 13:48:25.281460] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:28:18.481 [2024-11-20 13:48:25.281631] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:18.481 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:18.481 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:28:18.481 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:18.481 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:18.481 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:18.481 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:18.481 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:18.481 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:18.481 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:18.481 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:18.481 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:18.481 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:18.481 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:18.481 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:18.481 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:18.481 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:18.481 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:18.481 "name": "raid_bdev1", 00:28:18.481 "uuid": "23df9c16-cb60-4cc3-b79b-fca07d73faab", 00:28:18.481 "strip_size_kb": 64, 00:28:18.481 "state": "online", 00:28:18.481 "raid_level": "concat", 00:28:18.481 "superblock": true, 00:28:18.481 "num_base_bdevs": 3, 00:28:18.481 "num_base_bdevs_discovered": 3, 00:28:18.481 "num_base_bdevs_operational": 3, 00:28:18.481 "base_bdevs_list": [ 00:28:18.481 { 00:28:18.481 "name": "BaseBdev1", 00:28:18.482 "uuid": "f9b1db84-f890-5877-86d1-4756be21b084", 00:28:18.482 "is_configured": true, 00:28:18.482 "data_offset": 2048, 00:28:18.482 "data_size": 63488 00:28:18.482 }, 00:28:18.482 { 00:28:18.482 "name": "BaseBdev2", 00:28:18.482 "uuid": "72ae7bc7-eeb4-557f-bdee-960ab3ca1180", 00:28:18.482 "is_configured": true, 00:28:18.482 "data_offset": 2048, 00:28:18.482 "data_size": 63488 00:28:18.482 }, 00:28:18.482 { 00:28:18.482 "name": "BaseBdev3", 00:28:18.482 "uuid": "447f9ee7-ac5e-5770-8ae4-a74b6766a6dc", 00:28:18.482 "is_configured": true, 00:28:18.482 "data_offset": 2048, 00:28:18.482 "data_size": 63488 00:28:18.482 } 00:28:18.482 ] 00:28:18.482 }' 00:28:18.482 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:18.482 13:48:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:18.740 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:28:18.740 13:48:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:28:18.740 [2024-11-20 13:48:25.671541] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006700 00:28:19.673 13:48:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:28:19.673 13:48:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:19.673 13:48:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:19.673 13:48:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:19.673 13:48:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:28:19.673 13:48:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:28:19.673 13:48:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:28:19.673 13:48:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:28:19.673 13:48:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:19.673 13:48:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:19.673 13:48:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:19.673 13:48:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:19.673 13:48:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:19.673 13:48:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:19.673 13:48:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:19.673 13:48:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:19.673 13:48:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:19.673 13:48:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:19.673 13:48:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:19.673 13:48:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:19.673 13:48:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:19.673 13:48:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:19.673 13:48:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:19.673 "name": "raid_bdev1", 00:28:19.673 "uuid": "23df9c16-cb60-4cc3-b79b-fca07d73faab", 00:28:19.673 "strip_size_kb": 64, 00:28:19.673 "state": "online", 00:28:19.673 "raid_level": "concat", 00:28:19.673 "superblock": true, 00:28:19.673 "num_base_bdevs": 3, 00:28:19.673 "num_base_bdevs_discovered": 3, 00:28:19.673 "num_base_bdevs_operational": 3, 00:28:19.673 "base_bdevs_list": [ 00:28:19.673 { 00:28:19.673 "name": "BaseBdev1", 00:28:19.673 "uuid": "f9b1db84-f890-5877-86d1-4756be21b084", 00:28:19.673 "is_configured": true, 00:28:19.673 "data_offset": 2048, 00:28:19.673 "data_size": 63488 00:28:19.673 }, 00:28:19.673 { 00:28:19.673 "name": "BaseBdev2", 00:28:19.673 "uuid": "72ae7bc7-eeb4-557f-bdee-960ab3ca1180", 00:28:19.673 "is_configured": true, 00:28:19.673 "data_offset": 2048, 00:28:19.673 "data_size": 63488 00:28:19.673 }, 00:28:19.673 { 00:28:19.673 "name": "BaseBdev3", 00:28:19.673 "uuid": "447f9ee7-ac5e-5770-8ae4-a74b6766a6dc", 00:28:19.673 "is_configured": true, 00:28:19.673 "data_offset": 2048, 00:28:19.673 "data_size": 63488 00:28:19.673 } 00:28:19.673 ] 00:28:19.673 }' 00:28:19.673 13:48:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:19.673 13:48:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:19.931 13:48:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:28:19.931 13:48:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:19.931 13:48:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:19.931 [2024-11-20 13:48:26.893522] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:28:19.931 [2024-11-20 13:48:26.893565] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:28:19.931 [2024-11-20 13:48:26.896701] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:19.931 [2024-11-20 13:48:26.896759] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:19.931 [2024-11-20 13:48:26.896801] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:19.931 [2024-11-20 13:48:26.896813] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:28:19.931 13:48:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:19.931 13:48:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 65464 00:28:19.932 13:48:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 65464 ']' 00:28:19.932 13:48:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 65464 00:28:19.932 { 00:28:19.932 "results": [ 00:28:19.932 { 00:28:19.932 "job": "raid_bdev1", 00:28:19.932 "core_mask": "0x1", 00:28:19.932 "workload": "randrw", 00:28:19.932 "percentage": 50, 00:28:19.932 "status": "finished", 00:28:19.932 "queue_depth": 1, 00:28:19.932 "io_size": 131072, 00:28:19.932 "runtime": 1.22009, 00:28:19.932 "iops": 13555.557376914818, 00:28:19.932 "mibps": 1694.4446721143522, 00:28:19.932 "io_failed": 1, 00:28:19.932 "io_timeout": 0, 00:28:19.932 "avg_latency_us": 101.06530443679658, 00:28:19.932 "min_latency_us": 33.870769230769234, 00:28:19.932 "max_latency_us": 1701.4153846153847 00:28:19.932 } 00:28:19.932 ], 00:28:19.932 "core_count": 1 00:28:19.932 } 00:28:19.932 13:48:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:28:19.932 13:48:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:28:19.932 13:48:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 65464 00:28:19.932 13:48:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:28:19.932 13:48:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:28:19.932 killing process with pid 65464 00:28:19.932 13:48:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 65464' 00:28:19.932 13:48:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 65464 00:28:19.932 13:48:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 65464 00:28:19.932 [2024-11-20 13:48:26.924533] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:28:20.189 [2024-11-20 13:48:27.077231] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:28:21.123 13:48:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.kYptKtm8NQ 00:28:21.123 13:48:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:28:21.123 13:48:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:28:21.123 13:48:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.82 00:28:21.123 13:48:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:28:21.123 13:48:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:28:21.123 13:48:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:28:21.123 13:48:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.82 != \0\.\0\0 ]] 00:28:21.123 00:28:21.123 real 0m3.727s 00:28:21.123 user 0m4.351s 00:28:21.123 sys 0m0.442s 00:28:21.123 13:48:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:28:21.123 13:48:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:21.123 ************************************ 00:28:21.123 END TEST raid_read_error_test 00:28:21.123 ************************************ 00:28:21.123 13:48:27 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 3 write 00:28:21.123 13:48:27 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:28:21.123 13:48:27 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:28:21.123 13:48:27 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:28:21.123 ************************************ 00:28:21.123 START TEST raid_write_error_test 00:28:21.123 ************************************ 00:28:21.123 13:48:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 3 write 00:28:21.123 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:28:21.123 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:28:21.123 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:28:21.123 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:28:21.123 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:28:21.123 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:28:21.123 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:28:21.123 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:28:21.123 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:28:21.123 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:28:21.123 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:28:21.123 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:28:21.123 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:28:21.123 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:28:21.123 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:28:21.123 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:28:21.123 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:28:21.124 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:28:21.124 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:28:21.124 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:28:21.124 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:28:21.124 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:28:21.124 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:28:21.124 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:28:21.124 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:28:21.124 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.VB0m1udu51 00:28:21.124 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=65600 00:28:21.124 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 65600 00:28:21.124 13:48:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 65600 ']' 00:28:21.124 13:48:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:28:21.124 13:48:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:28:21.124 13:48:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:28:21.124 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:28:21.124 13:48:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:28:21.124 13:48:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:28:21.124 13:48:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:21.124 [2024-11-20 13:48:28.026804] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:28:21.124 [2024-11-20 13:48:28.027015] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid65600 ] 00:28:21.383 [2024-11-20 13:48:28.206205] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:21.383 [2024-11-20 13:48:28.372650] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:28:21.685 [2024-11-20 13:48:28.520986] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:21.685 [2024-11-20 13:48:28.521064] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:21.943 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:28:21.943 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:21.944 BaseBdev1_malloc 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:21.944 true 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:21.944 [2024-11-20 13:48:28.888315] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:28:21.944 [2024-11-20 13:48:28.888387] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:21.944 [2024-11-20 13:48:28.888413] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:28:21.944 [2024-11-20 13:48:28.888424] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:21.944 [2024-11-20 13:48:28.890770] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:21.944 [2024-11-20 13:48:28.890822] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:28:21.944 BaseBdev1 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:21.944 BaseBdev2_malloc 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:21.944 true 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:21.944 [2024-11-20 13:48:28.934502] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:28:21.944 [2024-11-20 13:48:28.934575] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:21.944 [2024-11-20 13:48:28.934594] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:28:21.944 [2024-11-20 13:48:28.934605] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:21.944 [2024-11-20 13:48:28.936968] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:21.944 [2024-11-20 13:48:28.937035] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:28:21.944 BaseBdev2 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:21.944 BaseBdev3_malloc 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:21.944 true 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:21.944 [2024-11-20 13:48:28.993898] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:28:21.944 [2024-11-20 13:48:28.993974] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:21.944 [2024-11-20 13:48:28.994008] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:28:21.944 [2024-11-20 13:48:28.994020] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:21.944 [2024-11-20 13:48:28.996435] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:21.944 [2024-11-20 13:48:28.996485] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:28:21.944 BaseBdev3 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:21.944 13:48:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:22.202 [2024-11-20 13:48:29.002007] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:22.202 [2024-11-20 13:48:29.004077] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:28:22.202 [2024-11-20 13:48:29.004171] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:28:22.202 [2024-11-20 13:48:29.004395] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:28:22.202 [2024-11-20 13:48:29.004413] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:28:22.202 [2024-11-20 13:48:29.004730] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006560 00:28:22.202 [2024-11-20 13:48:29.004905] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:28:22.202 [2024-11-20 13:48:29.004924] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:28:22.202 [2024-11-20 13:48:29.005117] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:22.202 13:48:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:22.202 13:48:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:28:22.202 13:48:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:22.202 13:48:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:22.202 13:48:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:22.202 13:48:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:22.202 13:48:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:22.202 13:48:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:22.202 13:48:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:22.202 13:48:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:22.202 13:48:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:22.202 13:48:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:22.202 13:48:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:22.202 13:48:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:22.202 13:48:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:22.202 13:48:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:22.202 13:48:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:22.202 "name": "raid_bdev1", 00:28:22.202 "uuid": "85c417b3-3363-483e-8c81-17731d154867", 00:28:22.202 "strip_size_kb": 64, 00:28:22.202 "state": "online", 00:28:22.202 "raid_level": "concat", 00:28:22.202 "superblock": true, 00:28:22.202 "num_base_bdevs": 3, 00:28:22.202 "num_base_bdevs_discovered": 3, 00:28:22.202 "num_base_bdevs_operational": 3, 00:28:22.202 "base_bdevs_list": [ 00:28:22.202 { 00:28:22.202 "name": "BaseBdev1", 00:28:22.202 "uuid": "5b39dc83-fff2-5600-9020-3edd896a3536", 00:28:22.202 "is_configured": true, 00:28:22.202 "data_offset": 2048, 00:28:22.202 "data_size": 63488 00:28:22.202 }, 00:28:22.202 { 00:28:22.202 "name": "BaseBdev2", 00:28:22.202 "uuid": "1cf797f2-a232-5d43-aadc-561f53ea80bf", 00:28:22.202 "is_configured": true, 00:28:22.202 "data_offset": 2048, 00:28:22.202 "data_size": 63488 00:28:22.202 }, 00:28:22.202 { 00:28:22.202 "name": "BaseBdev3", 00:28:22.202 "uuid": "94eabfb7-077e-56ac-9cdd-783c027df176", 00:28:22.202 "is_configured": true, 00:28:22.202 "data_offset": 2048, 00:28:22.202 "data_size": 63488 00:28:22.202 } 00:28:22.202 ] 00:28:22.202 }' 00:28:22.202 13:48:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:22.202 13:48:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:22.460 13:48:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:28:22.460 13:48:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:28:22.461 [2024-11-20 13:48:29.415087] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006700 00:28:23.395 13:48:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:28:23.395 13:48:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:23.395 13:48:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:23.395 13:48:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:23.395 13:48:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:28:23.395 13:48:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:28:23.395 13:48:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:28:23.395 13:48:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:28:23.395 13:48:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:23.395 13:48:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:23.395 13:48:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:28:23.395 13:48:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:23.395 13:48:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:23.395 13:48:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:23.395 13:48:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:23.395 13:48:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:23.395 13:48:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:23.395 13:48:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:23.395 13:48:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:23.395 13:48:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:23.395 13:48:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:23.395 13:48:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:23.395 13:48:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:23.395 "name": "raid_bdev1", 00:28:23.395 "uuid": "85c417b3-3363-483e-8c81-17731d154867", 00:28:23.395 "strip_size_kb": 64, 00:28:23.395 "state": "online", 00:28:23.395 "raid_level": "concat", 00:28:23.395 "superblock": true, 00:28:23.395 "num_base_bdevs": 3, 00:28:23.395 "num_base_bdevs_discovered": 3, 00:28:23.395 "num_base_bdevs_operational": 3, 00:28:23.396 "base_bdevs_list": [ 00:28:23.396 { 00:28:23.396 "name": "BaseBdev1", 00:28:23.396 "uuid": "5b39dc83-fff2-5600-9020-3edd896a3536", 00:28:23.396 "is_configured": true, 00:28:23.396 "data_offset": 2048, 00:28:23.396 "data_size": 63488 00:28:23.396 }, 00:28:23.396 { 00:28:23.396 "name": "BaseBdev2", 00:28:23.396 "uuid": "1cf797f2-a232-5d43-aadc-561f53ea80bf", 00:28:23.396 "is_configured": true, 00:28:23.396 "data_offset": 2048, 00:28:23.396 "data_size": 63488 00:28:23.396 }, 00:28:23.396 { 00:28:23.396 "name": "BaseBdev3", 00:28:23.396 "uuid": "94eabfb7-077e-56ac-9cdd-783c027df176", 00:28:23.396 "is_configured": true, 00:28:23.396 "data_offset": 2048, 00:28:23.396 "data_size": 63488 00:28:23.396 } 00:28:23.396 ] 00:28:23.396 }' 00:28:23.396 13:48:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:23.396 13:48:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:23.655 13:48:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:28:23.655 13:48:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:23.655 13:48:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:23.655 [2024-11-20 13:48:30.657500] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:28:23.655 [2024-11-20 13:48:30.657546] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:28:23.655 [2024-11-20 13:48:30.660754] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:23.655 [2024-11-20 13:48:30.660822] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:23.655 [2024-11-20 13:48:30.660867] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:23.655 [2024-11-20 13:48:30.660877] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:28:23.655 { 00:28:23.655 "results": [ 00:28:23.655 { 00:28:23.655 "job": "raid_bdev1", 00:28:23.655 "core_mask": "0x1", 00:28:23.655 "workload": "randrw", 00:28:23.655 "percentage": 50, 00:28:23.655 "status": "finished", 00:28:23.655 "queue_depth": 1, 00:28:23.655 "io_size": 131072, 00:28:23.655 "runtime": 1.240426, 00:28:23.655 "iops": 13980.680830617868, 00:28:23.655 "mibps": 1747.5851038272335, 00:28:23.655 "io_failed": 1, 00:28:23.655 "io_timeout": 0, 00:28:23.655 "avg_latency_us": 97.4867902368058, 00:28:23.655 "min_latency_us": 33.673846153846156, 00:28:23.655 "max_latency_us": 1688.8123076923077 00:28:23.655 } 00:28:23.655 ], 00:28:23.655 "core_count": 1 00:28:23.655 } 00:28:23.655 13:48:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:23.655 13:48:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 65600 00:28:23.655 13:48:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 65600 ']' 00:28:23.655 13:48:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 65600 00:28:23.655 13:48:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:28:23.655 13:48:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:28:23.655 13:48:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 65600 00:28:23.655 13:48:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:28:23.655 killing process with pid 65600 00:28:23.655 13:48:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:28:23.655 13:48:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 65600' 00:28:23.655 13:48:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 65600 00:28:23.655 [2024-11-20 13:48:30.690110] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:28:23.655 13:48:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 65600 00:28:23.914 [2024-11-20 13:48:30.841739] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:28:24.847 13:48:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:28:24.847 13:48:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.VB0m1udu51 00:28:24.847 13:48:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:28:24.847 13:48:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.81 00:28:24.847 13:48:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:28:24.847 13:48:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:28:24.847 13:48:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:28:24.847 13:48:31 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.81 != \0\.\0\0 ]] 00:28:24.847 00:28:24.847 real 0m3.732s 00:28:24.847 user 0m4.351s 00:28:24.847 sys 0m0.461s 00:28:24.847 13:48:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:28:24.847 13:48:31 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:24.847 ************************************ 00:28:24.847 END TEST raid_write_error_test 00:28:24.847 ************************************ 00:28:24.847 13:48:31 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:28:24.847 13:48:31 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 3 false 00:28:24.847 13:48:31 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:28:24.847 13:48:31 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:28:24.847 13:48:31 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:28:24.847 ************************************ 00:28:24.847 START TEST raid_state_function_test 00:28:24.847 ************************************ 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 3 false 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=65732 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 65732' 00:28:24.847 Process raid pid: 65732 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 65732 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 65732 ']' 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:28:24.847 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:28:24.847 13:48:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:24.847 [2024-11-20 13:48:31.771818] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:28:24.847 [2024-11-20 13:48:31.771946] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:28:25.105 [2024-11-20 13:48:31.932210] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:25.105 [2024-11-20 13:48:32.051767] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:28:25.363 [2024-11-20 13:48:32.203238] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:25.363 [2024-11-20 13:48:32.203304] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:25.622 13:48:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:28:25.622 13:48:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:28:25.622 13:48:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:28:25.622 13:48:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:25.623 13:48:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:25.623 [2024-11-20 13:48:32.660841] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:28:25.623 [2024-11-20 13:48:32.660909] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:28:25.623 [2024-11-20 13:48:32.660921] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:28:25.623 [2024-11-20 13:48:32.660932] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:28:25.623 [2024-11-20 13:48:32.660938] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:28:25.623 [2024-11-20 13:48:32.660947] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:28:25.623 13:48:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:25.623 13:48:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:28:25.623 13:48:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:25.623 13:48:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:25.623 13:48:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:25.623 13:48:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:25.623 13:48:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:25.623 13:48:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:25.623 13:48:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:25.623 13:48:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:25.623 13:48:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:25.623 13:48:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:25.623 13:48:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:25.623 13:48:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:25.623 13:48:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:25.881 13:48:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:25.882 13:48:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:25.882 "name": "Existed_Raid", 00:28:25.882 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:25.882 "strip_size_kb": 0, 00:28:25.882 "state": "configuring", 00:28:25.882 "raid_level": "raid1", 00:28:25.882 "superblock": false, 00:28:25.882 "num_base_bdevs": 3, 00:28:25.882 "num_base_bdevs_discovered": 0, 00:28:25.882 "num_base_bdevs_operational": 3, 00:28:25.882 "base_bdevs_list": [ 00:28:25.882 { 00:28:25.882 "name": "BaseBdev1", 00:28:25.882 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:25.882 "is_configured": false, 00:28:25.882 "data_offset": 0, 00:28:25.882 "data_size": 0 00:28:25.882 }, 00:28:25.882 { 00:28:25.882 "name": "BaseBdev2", 00:28:25.882 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:25.882 "is_configured": false, 00:28:25.882 "data_offset": 0, 00:28:25.882 "data_size": 0 00:28:25.882 }, 00:28:25.882 { 00:28:25.882 "name": "BaseBdev3", 00:28:25.882 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:25.882 "is_configured": false, 00:28:25.882 "data_offset": 0, 00:28:25.882 "data_size": 0 00:28:25.882 } 00:28:25.882 ] 00:28:25.882 }' 00:28:25.882 13:48:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:25.882 13:48:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:26.141 13:48:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:28:26.141 13:48:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:26.141 13:48:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:26.141 [2024-11-20 13:48:32.996877] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:28:26.141 [2024-11-20 13:48:32.996932] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:26.141 [2024-11-20 13:48:33.004872] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:28:26.141 [2024-11-20 13:48:33.004935] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:28:26.141 [2024-11-20 13:48:33.004945] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:28:26.141 [2024-11-20 13:48:33.004955] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:28:26.141 [2024-11-20 13:48:33.004961] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:28:26.141 [2024-11-20 13:48:33.004971] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:26.141 [2024-11-20 13:48:33.039646] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:26.141 BaseBdev1 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:26.141 [ 00:28:26.141 { 00:28:26.141 "name": "BaseBdev1", 00:28:26.141 "aliases": [ 00:28:26.141 "2ffa3b8d-13ca-4c2d-8376-cd2fd2cc41ec" 00:28:26.141 ], 00:28:26.141 "product_name": "Malloc disk", 00:28:26.141 "block_size": 512, 00:28:26.141 "num_blocks": 65536, 00:28:26.141 "uuid": "2ffa3b8d-13ca-4c2d-8376-cd2fd2cc41ec", 00:28:26.141 "assigned_rate_limits": { 00:28:26.141 "rw_ios_per_sec": 0, 00:28:26.141 "rw_mbytes_per_sec": 0, 00:28:26.141 "r_mbytes_per_sec": 0, 00:28:26.141 "w_mbytes_per_sec": 0 00:28:26.141 }, 00:28:26.141 "claimed": true, 00:28:26.141 "claim_type": "exclusive_write", 00:28:26.141 "zoned": false, 00:28:26.141 "supported_io_types": { 00:28:26.141 "read": true, 00:28:26.141 "write": true, 00:28:26.141 "unmap": true, 00:28:26.141 "flush": true, 00:28:26.141 "reset": true, 00:28:26.141 "nvme_admin": false, 00:28:26.141 "nvme_io": false, 00:28:26.141 "nvme_io_md": false, 00:28:26.141 "write_zeroes": true, 00:28:26.141 "zcopy": true, 00:28:26.141 "get_zone_info": false, 00:28:26.141 "zone_management": false, 00:28:26.141 "zone_append": false, 00:28:26.141 "compare": false, 00:28:26.141 "compare_and_write": false, 00:28:26.141 "abort": true, 00:28:26.141 "seek_hole": false, 00:28:26.141 "seek_data": false, 00:28:26.141 "copy": true, 00:28:26.141 "nvme_iov_md": false 00:28:26.141 }, 00:28:26.141 "memory_domains": [ 00:28:26.141 { 00:28:26.141 "dma_device_id": "system", 00:28:26.141 "dma_device_type": 1 00:28:26.141 }, 00:28:26.141 { 00:28:26.141 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:26.141 "dma_device_type": 2 00:28:26.141 } 00:28:26.141 ], 00:28:26.141 "driver_specific": {} 00:28:26.141 } 00:28:26.141 ] 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:26.141 "name": "Existed_Raid", 00:28:26.141 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:26.141 "strip_size_kb": 0, 00:28:26.141 "state": "configuring", 00:28:26.141 "raid_level": "raid1", 00:28:26.141 "superblock": false, 00:28:26.141 "num_base_bdevs": 3, 00:28:26.141 "num_base_bdevs_discovered": 1, 00:28:26.141 "num_base_bdevs_operational": 3, 00:28:26.141 "base_bdevs_list": [ 00:28:26.141 { 00:28:26.141 "name": "BaseBdev1", 00:28:26.141 "uuid": "2ffa3b8d-13ca-4c2d-8376-cd2fd2cc41ec", 00:28:26.141 "is_configured": true, 00:28:26.141 "data_offset": 0, 00:28:26.141 "data_size": 65536 00:28:26.141 }, 00:28:26.141 { 00:28:26.141 "name": "BaseBdev2", 00:28:26.141 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:26.141 "is_configured": false, 00:28:26.141 "data_offset": 0, 00:28:26.141 "data_size": 0 00:28:26.141 }, 00:28:26.141 { 00:28:26.141 "name": "BaseBdev3", 00:28:26.141 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:26.141 "is_configured": false, 00:28:26.141 "data_offset": 0, 00:28:26.141 "data_size": 0 00:28:26.141 } 00:28:26.141 ] 00:28:26.141 }' 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:26.141 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:26.400 [2024-11-20 13:48:33.383852] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:28:26.400 [2024-11-20 13:48:33.383946] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:26.400 [2024-11-20 13:48:33.391932] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:26.400 [2024-11-20 13:48:33.394656] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:28:26.400 [2024-11-20 13:48:33.394733] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:28:26.400 [2024-11-20 13:48:33.394752] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:28:26.400 [2024-11-20 13:48:33.394768] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:26.400 "name": "Existed_Raid", 00:28:26.400 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:26.400 "strip_size_kb": 0, 00:28:26.400 "state": "configuring", 00:28:26.400 "raid_level": "raid1", 00:28:26.400 "superblock": false, 00:28:26.400 "num_base_bdevs": 3, 00:28:26.400 "num_base_bdevs_discovered": 1, 00:28:26.400 "num_base_bdevs_operational": 3, 00:28:26.400 "base_bdevs_list": [ 00:28:26.400 { 00:28:26.400 "name": "BaseBdev1", 00:28:26.400 "uuid": "2ffa3b8d-13ca-4c2d-8376-cd2fd2cc41ec", 00:28:26.400 "is_configured": true, 00:28:26.400 "data_offset": 0, 00:28:26.400 "data_size": 65536 00:28:26.400 }, 00:28:26.400 { 00:28:26.400 "name": "BaseBdev2", 00:28:26.400 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:26.400 "is_configured": false, 00:28:26.400 "data_offset": 0, 00:28:26.400 "data_size": 0 00:28:26.400 }, 00:28:26.400 { 00:28:26.400 "name": "BaseBdev3", 00:28:26.400 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:26.400 "is_configured": false, 00:28:26.400 "data_offset": 0, 00:28:26.400 "data_size": 0 00:28:26.400 } 00:28:26.400 ] 00:28:26.400 }' 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:26.400 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:26.657 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:28:26.657 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:26.657 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:26.915 [2024-11-20 13:48:33.728711] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:28:26.915 BaseBdev2 00:28:26.915 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:26.915 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:28:26.915 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:28:26.915 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:26.915 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:28:26.915 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:26.915 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:26.915 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:26.916 [ 00:28:26.916 { 00:28:26.916 "name": "BaseBdev2", 00:28:26.916 "aliases": [ 00:28:26.916 "270aa16c-6a91-49ef-a1a5-e92d84e12fde" 00:28:26.916 ], 00:28:26.916 "product_name": "Malloc disk", 00:28:26.916 "block_size": 512, 00:28:26.916 "num_blocks": 65536, 00:28:26.916 "uuid": "270aa16c-6a91-49ef-a1a5-e92d84e12fde", 00:28:26.916 "assigned_rate_limits": { 00:28:26.916 "rw_ios_per_sec": 0, 00:28:26.916 "rw_mbytes_per_sec": 0, 00:28:26.916 "r_mbytes_per_sec": 0, 00:28:26.916 "w_mbytes_per_sec": 0 00:28:26.916 }, 00:28:26.916 "claimed": true, 00:28:26.916 "claim_type": "exclusive_write", 00:28:26.916 "zoned": false, 00:28:26.916 "supported_io_types": { 00:28:26.916 "read": true, 00:28:26.916 "write": true, 00:28:26.916 "unmap": true, 00:28:26.916 "flush": true, 00:28:26.916 "reset": true, 00:28:26.916 "nvme_admin": false, 00:28:26.916 "nvme_io": false, 00:28:26.916 "nvme_io_md": false, 00:28:26.916 "write_zeroes": true, 00:28:26.916 "zcopy": true, 00:28:26.916 "get_zone_info": false, 00:28:26.916 "zone_management": false, 00:28:26.916 "zone_append": false, 00:28:26.916 "compare": false, 00:28:26.916 "compare_and_write": false, 00:28:26.916 "abort": true, 00:28:26.916 "seek_hole": false, 00:28:26.916 "seek_data": false, 00:28:26.916 "copy": true, 00:28:26.916 "nvme_iov_md": false 00:28:26.916 }, 00:28:26.916 "memory_domains": [ 00:28:26.916 { 00:28:26.916 "dma_device_id": "system", 00:28:26.916 "dma_device_type": 1 00:28:26.916 }, 00:28:26.916 { 00:28:26.916 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:26.916 "dma_device_type": 2 00:28:26.916 } 00:28:26.916 ], 00:28:26.916 "driver_specific": {} 00:28:26.916 } 00:28:26.916 ] 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:26.916 "name": "Existed_Raid", 00:28:26.916 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:26.916 "strip_size_kb": 0, 00:28:26.916 "state": "configuring", 00:28:26.916 "raid_level": "raid1", 00:28:26.916 "superblock": false, 00:28:26.916 "num_base_bdevs": 3, 00:28:26.916 "num_base_bdevs_discovered": 2, 00:28:26.916 "num_base_bdevs_operational": 3, 00:28:26.916 "base_bdevs_list": [ 00:28:26.916 { 00:28:26.916 "name": "BaseBdev1", 00:28:26.916 "uuid": "2ffa3b8d-13ca-4c2d-8376-cd2fd2cc41ec", 00:28:26.916 "is_configured": true, 00:28:26.916 "data_offset": 0, 00:28:26.916 "data_size": 65536 00:28:26.916 }, 00:28:26.916 { 00:28:26.916 "name": "BaseBdev2", 00:28:26.916 "uuid": "270aa16c-6a91-49ef-a1a5-e92d84e12fde", 00:28:26.916 "is_configured": true, 00:28:26.916 "data_offset": 0, 00:28:26.916 "data_size": 65536 00:28:26.916 }, 00:28:26.916 { 00:28:26.916 "name": "BaseBdev3", 00:28:26.916 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:26.916 "is_configured": false, 00:28:26.916 "data_offset": 0, 00:28:26.916 "data_size": 0 00:28:26.916 } 00:28:26.916 ] 00:28:26.916 }' 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:26.916 13:48:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:27.174 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:28:27.174 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:27.174 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:27.174 [2024-11-20 13:48:34.145250] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:28:27.174 [2024-11-20 13:48:34.145313] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:28:27.174 [2024-11-20 13:48:34.145327] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:28:27.174 [2024-11-20 13:48:34.145621] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:28:27.174 [2024-11-20 13:48:34.145803] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:28:27.174 [2024-11-20 13:48:34.145818] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:28:27.174 [2024-11-20 13:48:34.146117] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:27.174 BaseBdev3 00:28:27.174 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:27.174 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:28:27.174 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:28:27.174 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:27.174 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:28:27.174 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:27.174 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:27.174 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:27.174 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:27.174 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:27.174 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:27.174 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:28:27.174 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:27.174 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:27.174 [ 00:28:27.174 { 00:28:27.174 "name": "BaseBdev3", 00:28:27.174 "aliases": [ 00:28:27.174 "f7f33a15-be65-4f34-bdd3-3794db2755c9" 00:28:27.174 ], 00:28:27.174 "product_name": "Malloc disk", 00:28:27.174 "block_size": 512, 00:28:27.175 "num_blocks": 65536, 00:28:27.175 "uuid": "f7f33a15-be65-4f34-bdd3-3794db2755c9", 00:28:27.175 "assigned_rate_limits": { 00:28:27.175 "rw_ios_per_sec": 0, 00:28:27.175 "rw_mbytes_per_sec": 0, 00:28:27.175 "r_mbytes_per_sec": 0, 00:28:27.175 "w_mbytes_per_sec": 0 00:28:27.175 }, 00:28:27.175 "claimed": true, 00:28:27.175 "claim_type": "exclusive_write", 00:28:27.175 "zoned": false, 00:28:27.175 "supported_io_types": { 00:28:27.175 "read": true, 00:28:27.175 "write": true, 00:28:27.175 "unmap": true, 00:28:27.175 "flush": true, 00:28:27.175 "reset": true, 00:28:27.175 "nvme_admin": false, 00:28:27.175 "nvme_io": false, 00:28:27.175 "nvme_io_md": false, 00:28:27.175 "write_zeroes": true, 00:28:27.175 "zcopy": true, 00:28:27.175 "get_zone_info": false, 00:28:27.175 "zone_management": false, 00:28:27.175 "zone_append": false, 00:28:27.175 "compare": false, 00:28:27.175 "compare_and_write": false, 00:28:27.175 "abort": true, 00:28:27.175 "seek_hole": false, 00:28:27.175 "seek_data": false, 00:28:27.175 "copy": true, 00:28:27.175 "nvme_iov_md": false 00:28:27.175 }, 00:28:27.175 "memory_domains": [ 00:28:27.175 { 00:28:27.175 "dma_device_id": "system", 00:28:27.175 "dma_device_type": 1 00:28:27.175 }, 00:28:27.175 { 00:28:27.175 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:27.175 "dma_device_type": 2 00:28:27.175 } 00:28:27.175 ], 00:28:27.175 "driver_specific": {} 00:28:27.175 } 00:28:27.175 ] 00:28:27.175 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:27.175 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:28:27.175 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:28:27.175 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:28:27.175 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:28:27.175 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:27.175 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:27.175 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:27.175 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:27.175 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:27.175 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:27.175 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:27.175 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:27.175 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:27.175 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:27.175 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:27.175 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:27.175 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:27.175 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:27.175 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:27.175 "name": "Existed_Raid", 00:28:27.175 "uuid": "173341c7-4b0e-4491-a890-3761ecfaeb23", 00:28:27.175 "strip_size_kb": 0, 00:28:27.175 "state": "online", 00:28:27.175 "raid_level": "raid1", 00:28:27.175 "superblock": false, 00:28:27.175 "num_base_bdevs": 3, 00:28:27.175 "num_base_bdevs_discovered": 3, 00:28:27.175 "num_base_bdevs_operational": 3, 00:28:27.175 "base_bdevs_list": [ 00:28:27.175 { 00:28:27.175 "name": "BaseBdev1", 00:28:27.175 "uuid": "2ffa3b8d-13ca-4c2d-8376-cd2fd2cc41ec", 00:28:27.175 "is_configured": true, 00:28:27.175 "data_offset": 0, 00:28:27.175 "data_size": 65536 00:28:27.175 }, 00:28:27.175 { 00:28:27.175 "name": "BaseBdev2", 00:28:27.175 "uuid": "270aa16c-6a91-49ef-a1a5-e92d84e12fde", 00:28:27.175 "is_configured": true, 00:28:27.175 "data_offset": 0, 00:28:27.175 "data_size": 65536 00:28:27.175 }, 00:28:27.175 { 00:28:27.175 "name": "BaseBdev3", 00:28:27.175 "uuid": "f7f33a15-be65-4f34-bdd3-3794db2755c9", 00:28:27.175 "is_configured": true, 00:28:27.175 "data_offset": 0, 00:28:27.175 "data_size": 65536 00:28:27.175 } 00:28:27.175 ] 00:28:27.175 }' 00:28:27.175 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:27.175 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:27.433 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:28:27.433 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:28:27.433 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:28:27.434 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:28:27.434 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:28:27.434 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:28:27.434 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:28:27.434 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:28:27.434 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:27.434 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:27.434 [2024-11-20 13:48:34.489764] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:27.692 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:27.692 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:27.692 "name": "Existed_Raid", 00:28:27.692 "aliases": [ 00:28:27.692 "173341c7-4b0e-4491-a890-3761ecfaeb23" 00:28:27.692 ], 00:28:27.692 "product_name": "Raid Volume", 00:28:27.692 "block_size": 512, 00:28:27.692 "num_blocks": 65536, 00:28:27.692 "uuid": "173341c7-4b0e-4491-a890-3761ecfaeb23", 00:28:27.692 "assigned_rate_limits": { 00:28:27.692 "rw_ios_per_sec": 0, 00:28:27.692 "rw_mbytes_per_sec": 0, 00:28:27.692 "r_mbytes_per_sec": 0, 00:28:27.692 "w_mbytes_per_sec": 0 00:28:27.692 }, 00:28:27.692 "claimed": false, 00:28:27.692 "zoned": false, 00:28:27.692 "supported_io_types": { 00:28:27.692 "read": true, 00:28:27.692 "write": true, 00:28:27.692 "unmap": false, 00:28:27.692 "flush": false, 00:28:27.692 "reset": true, 00:28:27.692 "nvme_admin": false, 00:28:27.692 "nvme_io": false, 00:28:27.692 "nvme_io_md": false, 00:28:27.692 "write_zeroes": true, 00:28:27.692 "zcopy": false, 00:28:27.692 "get_zone_info": false, 00:28:27.692 "zone_management": false, 00:28:27.692 "zone_append": false, 00:28:27.692 "compare": false, 00:28:27.692 "compare_and_write": false, 00:28:27.692 "abort": false, 00:28:27.692 "seek_hole": false, 00:28:27.692 "seek_data": false, 00:28:27.692 "copy": false, 00:28:27.692 "nvme_iov_md": false 00:28:27.692 }, 00:28:27.692 "memory_domains": [ 00:28:27.692 { 00:28:27.692 "dma_device_id": "system", 00:28:27.692 "dma_device_type": 1 00:28:27.692 }, 00:28:27.692 { 00:28:27.692 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:27.692 "dma_device_type": 2 00:28:27.692 }, 00:28:27.692 { 00:28:27.692 "dma_device_id": "system", 00:28:27.692 "dma_device_type": 1 00:28:27.692 }, 00:28:27.692 { 00:28:27.692 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:27.692 "dma_device_type": 2 00:28:27.692 }, 00:28:27.692 { 00:28:27.692 "dma_device_id": "system", 00:28:27.692 "dma_device_type": 1 00:28:27.692 }, 00:28:27.692 { 00:28:27.692 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:27.692 "dma_device_type": 2 00:28:27.692 } 00:28:27.692 ], 00:28:27.692 "driver_specific": { 00:28:27.692 "raid": { 00:28:27.692 "uuid": "173341c7-4b0e-4491-a890-3761ecfaeb23", 00:28:27.692 "strip_size_kb": 0, 00:28:27.692 "state": "online", 00:28:27.692 "raid_level": "raid1", 00:28:27.692 "superblock": false, 00:28:27.692 "num_base_bdevs": 3, 00:28:27.692 "num_base_bdevs_discovered": 3, 00:28:27.692 "num_base_bdevs_operational": 3, 00:28:27.692 "base_bdevs_list": [ 00:28:27.692 { 00:28:27.692 "name": "BaseBdev1", 00:28:27.692 "uuid": "2ffa3b8d-13ca-4c2d-8376-cd2fd2cc41ec", 00:28:27.692 "is_configured": true, 00:28:27.692 "data_offset": 0, 00:28:27.692 "data_size": 65536 00:28:27.692 }, 00:28:27.692 { 00:28:27.692 "name": "BaseBdev2", 00:28:27.692 "uuid": "270aa16c-6a91-49ef-a1a5-e92d84e12fde", 00:28:27.692 "is_configured": true, 00:28:27.692 "data_offset": 0, 00:28:27.692 "data_size": 65536 00:28:27.692 }, 00:28:27.692 { 00:28:27.692 "name": "BaseBdev3", 00:28:27.692 "uuid": "f7f33a15-be65-4f34-bdd3-3794db2755c9", 00:28:27.692 "is_configured": true, 00:28:27.692 "data_offset": 0, 00:28:27.692 "data_size": 65536 00:28:27.692 } 00:28:27.692 ] 00:28:27.692 } 00:28:27.693 } 00:28:27.693 }' 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:28:27.693 BaseBdev2 00:28:27.693 BaseBdev3' 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:27.693 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:27.693 [2024-11-20 13:48:34.689523] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:28:27.951 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:27.951 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:28:27.951 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:28:27.951 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:28:27.951 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:28:27.951 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:28:27.951 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:28:27.951 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:27.951 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:27.951 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:27.951 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:27.951 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:28:27.951 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:27.951 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:27.951 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:27.951 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:27.951 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:27.951 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:27.951 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:27.951 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:27.951 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:27.951 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:27.951 "name": "Existed_Raid", 00:28:27.951 "uuid": "173341c7-4b0e-4491-a890-3761ecfaeb23", 00:28:27.951 "strip_size_kb": 0, 00:28:27.951 "state": "online", 00:28:27.951 "raid_level": "raid1", 00:28:27.951 "superblock": false, 00:28:27.951 "num_base_bdevs": 3, 00:28:27.951 "num_base_bdevs_discovered": 2, 00:28:27.951 "num_base_bdevs_operational": 2, 00:28:27.951 "base_bdevs_list": [ 00:28:27.951 { 00:28:27.951 "name": null, 00:28:27.951 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:27.951 "is_configured": false, 00:28:27.951 "data_offset": 0, 00:28:27.951 "data_size": 65536 00:28:27.951 }, 00:28:27.951 { 00:28:27.951 "name": "BaseBdev2", 00:28:27.951 "uuid": "270aa16c-6a91-49ef-a1a5-e92d84e12fde", 00:28:27.951 "is_configured": true, 00:28:27.951 "data_offset": 0, 00:28:27.951 "data_size": 65536 00:28:27.951 }, 00:28:27.951 { 00:28:27.951 "name": "BaseBdev3", 00:28:27.951 "uuid": "f7f33a15-be65-4f34-bdd3-3794db2755c9", 00:28:27.951 "is_configured": true, 00:28:27.951 "data_offset": 0, 00:28:27.951 "data_size": 65536 00:28:27.951 } 00:28:27.951 ] 00:28:27.951 }' 00:28:27.951 13:48:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:27.951 13:48:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:28.209 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:28:28.209 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:28:28.209 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:28.209 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:28:28.209 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:28.209 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:28.209 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:28.209 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:28:28.209 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:28:28.209 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:28:28.209 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:28.209 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:28.209 [2024-11-20 13:48:35.125523] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:28:28.209 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:28.209 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:28:28.209 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:28:28.209 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:28:28.209 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:28.209 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:28.209 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:28.209 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:28.209 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:28:28.209 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:28:28.209 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:28:28.209 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:28.209 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:28.209 [2024-11-20 13:48:35.230074] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:28:28.209 [2024-11-20 13:48:35.230193] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:28:28.468 [2024-11-20 13:48:35.293538] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:28.468 [2024-11-20 13:48:35.293603] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:28.468 [2024-11-20 13:48:35.293615] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:28:28.468 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:28.468 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:28:28.468 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:28:28.468 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:28.468 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:28.468 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:28.468 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:28:28.468 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:28.468 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:28:28.468 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:28:28.468 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:28:28.468 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:28:28.468 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:28:28.468 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:28.469 BaseBdev2 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:28.469 [ 00:28:28.469 { 00:28:28.469 "name": "BaseBdev2", 00:28:28.469 "aliases": [ 00:28:28.469 "875d72a0-bb69-4c8b-9edc-2c59ef06429a" 00:28:28.469 ], 00:28:28.469 "product_name": "Malloc disk", 00:28:28.469 "block_size": 512, 00:28:28.469 "num_blocks": 65536, 00:28:28.469 "uuid": "875d72a0-bb69-4c8b-9edc-2c59ef06429a", 00:28:28.469 "assigned_rate_limits": { 00:28:28.469 "rw_ios_per_sec": 0, 00:28:28.469 "rw_mbytes_per_sec": 0, 00:28:28.469 "r_mbytes_per_sec": 0, 00:28:28.469 "w_mbytes_per_sec": 0 00:28:28.469 }, 00:28:28.469 "claimed": false, 00:28:28.469 "zoned": false, 00:28:28.469 "supported_io_types": { 00:28:28.469 "read": true, 00:28:28.469 "write": true, 00:28:28.469 "unmap": true, 00:28:28.469 "flush": true, 00:28:28.469 "reset": true, 00:28:28.469 "nvme_admin": false, 00:28:28.469 "nvme_io": false, 00:28:28.469 "nvme_io_md": false, 00:28:28.469 "write_zeroes": true, 00:28:28.469 "zcopy": true, 00:28:28.469 "get_zone_info": false, 00:28:28.469 "zone_management": false, 00:28:28.469 "zone_append": false, 00:28:28.469 "compare": false, 00:28:28.469 "compare_and_write": false, 00:28:28.469 "abort": true, 00:28:28.469 "seek_hole": false, 00:28:28.469 "seek_data": false, 00:28:28.469 "copy": true, 00:28:28.469 "nvme_iov_md": false 00:28:28.469 }, 00:28:28.469 "memory_domains": [ 00:28:28.469 { 00:28:28.469 "dma_device_id": "system", 00:28:28.469 "dma_device_type": 1 00:28:28.469 }, 00:28:28.469 { 00:28:28.469 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:28.469 "dma_device_type": 2 00:28:28.469 } 00:28:28.469 ], 00:28:28.469 "driver_specific": {} 00:28:28.469 } 00:28:28.469 ] 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:28.469 BaseBdev3 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:28.469 [ 00:28:28.469 { 00:28:28.469 "name": "BaseBdev3", 00:28:28.469 "aliases": [ 00:28:28.469 "d827d362-f2f6-4fd5-8c35-2ac94076582a" 00:28:28.469 ], 00:28:28.469 "product_name": "Malloc disk", 00:28:28.469 "block_size": 512, 00:28:28.469 "num_blocks": 65536, 00:28:28.469 "uuid": "d827d362-f2f6-4fd5-8c35-2ac94076582a", 00:28:28.469 "assigned_rate_limits": { 00:28:28.469 "rw_ios_per_sec": 0, 00:28:28.469 "rw_mbytes_per_sec": 0, 00:28:28.469 "r_mbytes_per_sec": 0, 00:28:28.469 "w_mbytes_per_sec": 0 00:28:28.469 }, 00:28:28.469 "claimed": false, 00:28:28.469 "zoned": false, 00:28:28.469 "supported_io_types": { 00:28:28.469 "read": true, 00:28:28.469 "write": true, 00:28:28.469 "unmap": true, 00:28:28.469 "flush": true, 00:28:28.469 "reset": true, 00:28:28.469 "nvme_admin": false, 00:28:28.469 "nvme_io": false, 00:28:28.469 "nvme_io_md": false, 00:28:28.469 "write_zeroes": true, 00:28:28.469 "zcopy": true, 00:28:28.469 "get_zone_info": false, 00:28:28.469 "zone_management": false, 00:28:28.469 "zone_append": false, 00:28:28.469 "compare": false, 00:28:28.469 "compare_and_write": false, 00:28:28.469 "abort": true, 00:28:28.469 "seek_hole": false, 00:28:28.469 "seek_data": false, 00:28:28.469 "copy": true, 00:28:28.469 "nvme_iov_md": false 00:28:28.469 }, 00:28:28.469 "memory_domains": [ 00:28:28.469 { 00:28:28.469 "dma_device_id": "system", 00:28:28.469 "dma_device_type": 1 00:28:28.469 }, 00:28:28.469 { 00:28:28.469 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:28.469 "dma_device_type": 2 00:28:28.469 } 00:28:28.469 ], 00:28:28.469 "driver_specific": {} 00:28:28.469 } 00:28:28.469 ] 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:28:28.469 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:28:28.470 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:28.470 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:28.470 [2024-11-20 13:48:35.446545] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:28:28.470 [2024-11-20 13:48:35.446615] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:28:28.470 [2024-11-20 13:48:35.446640] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:28:28.470 [2024-11-20 13:48:35.448727] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:28:28.470 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:28.470 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:28:28.470 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:28.470 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:28.470 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:28.470 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:28.470 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:28.470 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:28.470 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:28.470 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:28.470 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:28.470 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:28.470 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:28.470 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:28.470 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:28.470 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:28.470 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:28.470 "name": "Existed_Raid", 00:28:28.470 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:28.470 "strip_size_kb": 0, 00:28:28.470 "state": "configuring", 00:28:28.470 "raid_level": "raid1", 00:28:28.470 "superblock": false, 00:28:28.470 "num_base_bdevs": 3, 00:28:28.470 "num_base_bdevs_discovered": 2, 00:28:28.470 "num_base_bdevs_operational": 3, 00:28:28.470 "base_bdevs_list": [ 00:28:28.470 { 00:28:28.470 "name": "BaseBdev1", 00:28:28.470 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:28.470 "is_configured": false, 00:28:28.470 "data_offset": 0, 00:28:28.470 "data_size": 0 00:28:28.470 }, 00:28:28.470 { 00:28:28.470 "name": "BaseBdev2", 00:28:28.470 "uuid": "875d72a0-bb69-4c8b-9edc-2c59ef06429a", 00:28:28.470 "is_configured": true, 00:28:28.470 "data_offset": 0, 00:28:28.470 "data_size": 65536 00:28:28.470 }, 00:28:28.470 { 00:28:28.470 "name": "BaseBdev3", 00:28:28.470 "uuid": "d827d362-f2f6-4fd5-8c35-2ac94076582a", 00:28:28.470 "is_configured": true, 00:28:28.470 "data_offset": 0, 00:28:28.470 "data_size": 65536 00:28:28.470 } 00:28:28.470 ] 00:28:28.470 }' 00:28:28.470 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:28.470 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:28.728 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:28:28.728 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:28.728 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:28.728 [2024-11-20 13:48:35.782619] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:28:28.986 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:28.986 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:28:28.986 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:28.986 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:28.986 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:28.986 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:28.986 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:28.986 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:28.986 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:28.986 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:28.986 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:28.986 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:28.986 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:28.986 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:28.986 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:28.986 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:28.986 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:28.986 "name": "Existed_Raid", 00:28:28.986 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:28.986 "strip_size_kb": 0, 00:28:28.986 "state": "configuring", 00:28:28.986 "raid_level": "raid1", 00:28:28.986 "superblock": false, 00:28:28.986 "num_base_bdevs": 3, 00:28:28.986 "num_base_bdevs_discovered": 1, 00:28:28.986 "num_base_bdevs_operational": 3, 00:28:28.986 "base_bdevs_list": [ 00:28:28.986 { 00:28:28.986 "name": "BaseBdev1", 00:28:28.986 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:28.986 "is_configured": false, 00:28:28.986 "data_offset": 0, 00:28:28.986 "data_size": 0 00:28:28.986 }, 00:28:28.986 { 00:28:28.986 "name": null, 00:28:28.986 "uuid": "875d72a0-bb69-4c8b-9edc-2c59ef06429a", 00:28:28.986 "is_configured": false, 00:28:28.986 "data_offset": 0, 00:28:28.986 "data_size": 65536 00:28:28.986 }, 00:28:28.986 { 00:28:28.986 "name": "BaseBdev3", 00:28:28.986 "uuid": "d827d362-f2f6-4fd5-8c35-2ac94076582a", 00:28:28.986 "is_configured": true, 00:28:28.986 "data_offset": 0, 00:28:28.986 "data_size": 65536 00:28:28.986 } 00:28:28.986 ] 00:28:28.986 }' 00:28:28.986 13:48:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:28.986 13:48:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:29.244 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:29.244 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:29.244 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:29.244 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:28:29.244 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:29.244 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:28:29.244 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:28:29.244 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:29.244 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:29.244 [2024-11-20 13:48:36.179381] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:29.244 BaseBdev1 00:28:29.244 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:29.244 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:28:29.244 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:28:29.244 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:29.244 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:28:29.244 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:29.244 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:29.244 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:29.244 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:29.244 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:29.244 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:29.244 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:28:29.244 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:29.244 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:29.244 [ 00:28:29.244 { 00:28:29.245 "name": "BaseBdev1", 00:28:29.245 "aliases": [ 00:28:29.245 "b78aca2c-72ff-4759-b4f7-89fe90addb82" 00:28:29.245 ], 00:28:29.245 "product_name": "Malloc disk", 00:28:29.245 "block_size": 512, 00:28:29.245 "num_blocks": 65536, 00:28:29.245 "uuid": "b78aca2c-72ff-4759-b4f7-89fe90addb82", 00:28:29.245 "assigned_rate_limits": { 00:28:29.245 "rw_ios_per_sec": 0, 00:28:29.245 "rw_mbytes_per_sec": 0, 00:28:29.245 "r_mbytes_per_sec": 0, 00:28:29.245 "w_mbytes_per_sec": 0 00:28:29.245 }, 00:28:29.245 "claimed": true, 00:28:29.245 "claim_type": "exclusive_write", 00:28:29.245 "zoned": false, 00:28:29.245 "supported_io_types": { 00:28:29.245 "read": true, 00:28:29.245 "write": true, 00:28:29.245 "unmap": true, 00:28:29.245 "flush": true, 00:28:29.245 "reset": true, 00:28:29.245 "nvme_admin": false, 00:28:29.245 "nvme_io": false, 00:28:29.245 "nvme_io_md": false, 00:28:29.245 "write_zeroes": true, 00:28:29.245 "zcopy": true, 00:28:29.245 "get_zone_info": false, 00:28:29.245 "zone_management": false, 00:28:29.245 "zone_append": false, 00:28:29.245 "compare": false, 00:28:29.245 "compare_and_write": false, 00:28:29.245 "abort": true, 00:28:29.245 "seek_hole": false, 00:28:29.245 "seek_data": false, 00:28:29.245 "copy": true, 00:28:29.245 "nvme_iov_md": false 00:28:29.245 }, 00:28:29.245 "memory_domains": [ 00:28:29.245 { 00:28:29.245 "dma_device_id": "system", 00:28:29.245 "dma_device_type": 1 00:28:29.245 }, 00:28:29.245 { 00:28:29.245 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:29.245 "dma_device_type": 2 00:28:29.245 } 00:28:29.245 ], 00:28:29.245 "driver_specific": {} 00:28:29.245 } 00:28:29.245 ] 00:28:29.245 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:29.245 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:28:29.245 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:28:29.245 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:29.245 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:29.245 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:29.245 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:29.245 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:29.245 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:29.245 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:29.245 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:29.245 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:29.245 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:29.245 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:29.245 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:29.245 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:29.245 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:29.245 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:29.245 "name": "Existed_Raid", 00:28:29.245 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:29.245 "strip_size_kb": 0, 00:28:29.245 "state": "configuring", 00:28:29.245 "raid_level": "raid1", 00:28:29.245 "superblock": false, 00:28:29.245 "num_base_bdevs": 3, 00:28:29.245 "num_base_bdevs_discovered": 2, 00:28:29.245 "num_base_bdevs_operational": 3, 00:28:29.245 "base_bdevs_list": [ 00:28:29.245 { 00:28:29.245 "name": "BaseBdev1", 00:28:29.245 "uuid": "b78aca2c-72ff-4759-b4f7-89fe90addb82", 00:28:29.245 "is_configured": true, 00:28:29.245 "data_offset": 0, 00:28:29.245 "data_size": 65536 00:28:29.245 }, 00:28:29.245 { 00:28:29.245 "name": null, 00:28:29.245 "uuid": "875d72a0-bb69-4c8b-9edc-2c59ef06429a", 00:28:29.245 "is_configured": false, 00:28:29.245 "data_offset": 0, 00:28:29.245 "data_size": 65536 00:28:29.245 }, 00:28:29.245 { 00:28:29.245 "name": "BaseBdev3", 00:28:29.245 "uuid": "d827d362-f2f6-4fd5-8c35-2ac94076582a", 00:28:29.245 "is_configured": true, 00:28:29.245 "data_offset": 0, 00:28:29.245 "data_size": 65536 00:28:29.245 } 00:28:29.245 ] 00:28:29.245 }' 00:28:29.245 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:29.245 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:29.504 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:29.504 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:29.504 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:29.504 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:28:29.504 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:29.504 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:28:29.504 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:28:29.504 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:29.504 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:29.504 [2024-11-20 13:48:36.543548] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:28:29.504 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:29.504 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:28:29.504 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:29.504 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:29.504 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:29.504 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:29.504 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:29.504 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:29.504 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:29.504 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:29.504 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:29.504 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:29.504 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:29.504 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:29.504 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:29.762 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:29.762 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:29.762 "name": "Existed_Raid", 00:28:29.762 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:29.762 "strip_size_kb": 0, 00:28:29.762 "state": "configuring", 00:28:29.762 "raid_level": "raid1", 00:28:29.762 "superblock": false, 00:28:29.762 "num_base_bdevs": 3, 00:28:29.762 "num_base_bdevs_discovered": 1, 00:28:29.762 "num_base_bdevs_operational": 3, 00:28:29.762 "base_bdevs_list": [ 00:28:29.762 { 00:28:29.762 "name": "BaseBdev1", 00:28:29.762 "uuid": "b78aca2c-72ff-4759-b4f7-89fe90addb82", 00:28:29.762 "is_configured": true, 00:28:29.762 "data_offset": 0, 00:28:29.762 "data_size": 65536 00:28:29.762 }, 00:28:29.762 { 00:28:29.762 "name": null, 00:28:29.762 "uuid": "875d72a0-bb69-4c8b-9edc-2c59ef06429a", 00:28:29.762 "is_configured": false, 00:28:29.762 "data_offset": 0, 00:28:29.762 "data_size": 65536 00:28:29.762 }, 00:28:29.762 { 00:28:29.762 "name": null, 00:28:29.762 "uuid": "d827d362-f2f6-4fd5-8c35-2ac94076582a", 00:28:29.762 "is_configured": false, 00:28:29.762 "data_offset": 0, 00:28:29.762 "data_size": 65536 00:28:29.762 } 00:28:29.762 ] 00:28:29.762 }' 00:28:29.762 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:29.762 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:30.021 [2024-11-20 13:48:36.891692] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:30.021 "name": "Existed_Raid", 00:28:30.021 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:30.021 "strip_size_kb": 0, 00:28:30.021 "state": "configuring", 00:28:30.021 "raid_level": "raid1", 00:28:30.021 "superblock": false, 00:28:30.021 "num_base_bdevs": 3, 00:28:30.021 "num_base_bdevs_discovered": 2, 00:28:30.021 "num_base_bdevs_operational": 3, 00:28:30.021 "base_bdevs_list": [ 00:28:30.021 { 00:28:30.021 "name": "BaseBdev1", 00:28:30.021 "uuid": "b78aca2c-72ff-4759-b4f7-89fe90addb82", 00:28:30.021 "is_configured": true, 00:28:30.021 "data_offset": 0, 00:28:30.021 "data_size": 65536 00:28:30.021 }, 00:28:30.021 { 00:28:30.021 "name": null, 00:28:30.021 "uuid": "875d72a0-bb69-4c8b-9edc-2c59ef06429a", 00:28:30.021 "is_configured": false, 00:28:30.021 "data_offset": 0, 00:28:30.021 "data_size": 65536 00:28:30.021 }, 00:28:30.021 { 00:28:30.021 "name": "BaseBdev3", 00:28:30.021 "uuid": "d827d362-f2f6-4fd5-8c35-2ac94076582a", 00:28:30.021 "is_configured": true, 00:28:30.021 "data_offset": 0, 00:28:30.021 "data_size": 65536 00:28:30.021 } 00:28:30.021 ] 00:28:30.021 }' 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:30.021 13:48:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:30.280 [2024-11-20 13:48:37.223776] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:30.280 "name": "Existed_Raid", 00:28:30.280 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:30.280 "strip_size_kb": 0, 00:28:30.280 "state": "configuring", 00:28:30.280 "raid_level": "raid1", 00:28:30.280 "superblock": false, 00:28:30.280 "num_base_bdevs": 3, 00:28:30.280 "num_base_bdevs_discovered": 1, 00:28:30.280 "num_base_bdevs_operational": 3, 00:28:30.280 "base_bdevs_list": [ 00:28:30.280 { 00:28:30.280 "name": null, 00:28:30.280 "uuid": "b78aca2c-72ff-4759-b4f7-89fe90addb82", 00:28:30.280 "is_configured": false, 00:28:30.280 "data_offset": 0, 00:28:30.280 "data_size": 65536 00:28:30.280 }, 00:28:30.280 { 00:28:30.280 "name": null, 00:28:30.280 "uuid": "875d72a0-bb69-4c8b-9edc-2c59ef06429a", 00:28:30.280 "is_configured": false, 00:28:30.280 "data_offset": 0, 00:28:30.280 "data_size": 65536 00:28:30.280 }, 00:28:30.280 { 00:28:30.280 "name": "BaseBdev3", 00:28:30.280 "uuid": "d827d362-f2f6-4fd5-8c35-2ac94076582a", 00:28:30.280 "is_configured": true, 00:28:30.280 "data_offset": 0, 00:28:30.280 "data_size": 65536 00:28:30.280 } 00:28:30.280 ] 00:28:30.280 }' 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:30.280 13:48:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:30.927 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:28:30.927 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:30.927 13:48:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:30.927 13:48:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:30.927 13:48:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:30.927 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:28:30.927 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:28:30.927 13:48:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:30.927 13:48:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:30.927 [2024-11-20 13:48:37.674369] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:28:30.927 13:48:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:30.927 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:28:30.928 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:30.928 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:30.928 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:30.928 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:30.928 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:30.928 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:30.928 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:30.928 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:30.928 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:30.928 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:30.928 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:30.928 13:48:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:30.928 13:48:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:30.928 13:48:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:30.928 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:30.928 "name": "Existed_Raid", 00:28:30.928 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:30.928 "strip_size_kb": 0, 00:28:30.928 "state": "configuring", 00:28:30.928 "raid_level": "raid1", 00:28:30.928 "superblock": false, 00:28:30.928 "num_base_bdevs": 3, 00:28:30.928 "num_base_bdevs_discovered": 2, 00:28:30.928 "num_base_bdevs_operational": 3, 00:28:30.928 "base_bdevs_list": [ 00:28:30.928 { 00:28:30.928 "name": null, 00:28:30.928 "uuid": "b78aca2c-72ff-4759-b4f7-89fe90addb82", 00:28:30.928 "is_configured": false, 00:28:30.928 "data_offset": 0, 00:28:30.928 "data_size": 65536 00:28:30.928 }, 00:28:30.928 { 00:28:30.928 "name": "BaseBdev2", 00:28:30.928 "uuid": "875d72a0-bb69-4c8b-9edc-2c59ef06429a", 00:28:30.928 "is_configured": true, 00:28:30.928 "data_offset": 0, 00:28:30.928 "data_size": 65536 00:28:30.928 }, 00:28:30.928 { 00:28:30.928 "name": "BaseBdev3", 00:28:30.928 "uuid": "d827d362-f2f6-4fd5-8c35-2ac94076582a", 00:28:30.928 "is_configured": true, 00:28:30.928 "data_offset": 0, 00:28:30.928 "data_size": 65536 00:28:30.928 } 00:28:30.928 ] 00:28:30.928 }' 00:28:30.928 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:30.928 13:48:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:31.188 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:28:31.188 13:48:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:31.188 13:48:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:31.188 13:48:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:31.188 13:48:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:31.188 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:28:31.188 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:31.188 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:31.188 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:31.188 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:28:31.188 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:31.188 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u b78aca2c-72ff-4759-b4f7-89fe90addb82 00:28:31.188 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:31.188 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:31.188 [2024-11-20 13:48:38.063191] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:28:31.188 [2024-11-20 13:48:38.063245] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:28:31.188 [2024-11-20 13:48:38.063252] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:28:31.188 [2024-11-20 13:48:38.063497] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:28:31.188 [2024-11-20 13:48:38.063627] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:28:31.188 [2024-11-20 13:48:38.063638] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:28:31.188 [2024-11-20 13:48:38.063861] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:31.188 NewBaseBdev 00:28:31.188 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:31.188 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:28:31.188 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:31.189 [ 00:28:31.189 { 00:28:31.189 "name": "NewBaseBdev", 00:28:31.189 "aliases": [ 00:28:31.189 "b78aca2c-72ff-4759-b4f7-89fe90addb82" 00:28:31.189 ], 00:28:31.189 "product_name": "Malloc disk", 00:28:31.189 "block_size": 512, 00:28:31.189 "num_blocks": 65536, 00:28:31.189 "uuid": "b78aca2c-72ff-4759-b4f7-89fe90addb82", 00:28:31.189 "assigned_rate_limits": { 00:28:31.189 "rw_ios_per_sec": 0, 00:28:31.189 "rw_mbytes_per_sec": 0, 00:28:31.189 "r_mbytes_per_sec": 0, 00:28:31.189 "w_mbytes_per_sec": 0 00:28:31.189 }, 00:28:31.189 "claimed": true, 00:28:31.189 "claim_type": "exclusive_write", 00:28:31.189 "zoned": false, 00:28:31.189 "supported_io_types": { 00:28:31.189 "read": true, 00:28:31.189 "write": true, 00:28:31.189 "unmap": true, 00:28:31.189 "flush": true, 00:28:31.189 "reset": true, 00:28:31.189 "nvme_admin": false, 00:28:31.189 "nvme_io": false, 00:28:31.189 "nvme_io_md": false, 00:28:31.189 "write_zeroes": true, 00:28:31.189 "zcopy": true, 00:28:31.189 "get_zone_info": false, 00:28:31.189 "zone_management": false, 00:28:31.189 "zone_append": false, 00:28:31.189 "compare": false, 00:28:31.189 "compare_and_write": false, 00:28:31.189 "abort": true, 00:28:31.189 "seek_hole": false, 00:28:31.189 "seek_data": false, 00:28:31.189 "copy": true, 00:28:31.189 "nvme_iov_md": false 00:28:31.189 }, 00:28:31.189 "memory_domains": [ 00:28:31.189 { 00:28:31.189 "dma_device_id": "system", 00:28:31.189 "dma_device_type": 1 00:28:31.189 }, 00:28:31.189 { 00:28:31.189 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:31.189 "dma_device_type": 2 00:28:31.189 } 00:28:31.189 ], 00:28:31.189 "driver_specific": {} 00:28:31.189 } 00:28:31.189 ] 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:31.189 "name": "Existed_Raid", 00:28:31.189 "uuid": "0a49480c-abe0-4d7c-8b72-d1e96fa180d9", 00:28:31.189 "strip_size_kb": 0, 00:28:31.189 "state": "online", 00:28:31.189 "raid_level": "raid1", 00:28:31.189 "superblock": false, 00:28:31.189 "num_base_bdevs": 3, 00:28:31.189 "num_base_bdevs_discovered": 3, 00:28:31.189 "num_base_bdevs_operational": 3, 00:28:31.189 "base_bdevs_list": [ 00:28:31.189 { 00:28:31.189 "name": "NewBaseBdev", 00:28:31.189 "uuid": "b78aca2c-72ff-4759-b4f7-89fe90addb82", 00:28:31.189 "is_configured": true, 00:28:31.189 "data_offset": 0, 00:28:31.189 "data_size": 65536 00:28:31.189 }, 00:28:31.189 { 00:28:31.189 "name": "BaseBdev2", 00:28:31.189 "uuid": "875d72a0-bb69-4c8b-9edc-2c59ef06429a", 00:28:31.189 "is_configured": true, 00:28:31.189 "data_offset": 0, 00:28:31.189 "data_size": 65536 00:28:31.189 }, 00:28:31.189 { 00:28:31.189 "name": "BaseBdev3", 00:28:31.189 "uuid": "d827d362-f2f6-4fd5-8c35-2ac94076582a", 00:28:31.189 "is_configured": true, 00:28:31.189 "data_offset": 0, 00:28:31.189 "data_size": 65536 00:28:31.189 } 00:28:31.189 ] 00:28:31.189 }' 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:31.189 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:31.448 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:28:31.448 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:28:31.448 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:28:31.448 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:28:31.448 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:28:31.448 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:28:31.448 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:28:31.448 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:31.448 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:31.448 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:28:31.448 [2024-11-20 13:48:38.423626] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:31.448 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:31.448 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:31.448 "name": "Existed_Raid", 00:28:31.448 "aliases": [ 00:28:31.448 "0a49480c-abe0-4d7c-8b72-d1e96fa180d9" 00:28:31.448 ], 00:28:31.448 "product_name": "Raid Volume", 00:28:31.448 "block_size": 512, 00:28:31.448 "num_blocks": 65536, 00:28:31.448 "uuid": "0a49480c-abe0-4d7c-8b72-d1e96fa180d9", 00:28:31.448 "assigned_rate_limits": { 00:28:31.448 "rw_ios_per_sec": 0, 00:28:31.449 "rw_mbytes_per_sec": 0, 00:28:31.449 "r_mbytes_per_sec": 0, 00:28:31.449 "w_mbytes_per_sec": 0 00:28:31.449 }, 00:28:31.449 "claimed": false, 00:28:31.449 "zoned": false, 00:28:31.449 "supported_io_types": { 00:28:31.449 "read": true, 00:28:31.449 "write": true, 00:28:31.449 "unmap": false, 00:28:31.449 "flush": false, 00:28:31.449 "reset": true, 00:28:31.449 "nvme_admin": false, 00:28:31.449 "nvme_io": false, 00:28:31.449 "nvme_io_md": false, 00:28:31.449 "write_zeroes": true, 00:28:31.449 "zcopy": false, 00:28:31.449 "get_zone_info": false, 00:28:31.449 "zone_management": false, 00:28:31.449 "zone_append": false, 00:28:31.449 "compare": false, 00:28:31.449 "compare_and_write": false, 00:28:31.449 "abort": false, 00:28:31.449 "seek_hole": false, 00:28:31.449 "seek_data": false, 00:28:31.449 "copy": false, 00:28:31.449 "nvme_iov_md": false 00:28:31.449 }, 00:28:31.449 "memory_domains": [ 00:28:31.449 { 00:28:31.449 "dma_device_id": "system", 00:28:31.449 "dma_device_type": 1 00:28:31.449 }, 00:28:31.449 { 00:28:31.449 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:31.449 "dma_device_type": 2 00:28:31.449 }, 00:28:31.449 { 00:28:31.449 "dma_device_id": "system", 00:28:31.449 "dma_device_type": 1 00:28:31.449 }, 00:28:31.449 { 00:28:31.449 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:31.449 "dma_device_type": 2 00:28:31.449 }, 00:28:31.449 { 00:28:31.449 "dma_device_id": "system", 00:28:31.449 "dma_device_type": 1 00:28:31.449 }, 00:28:31.449 { 00:28:31.449 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:31.449 "dma_device_type": 2 00:28:31.449 } 00:28:31.449 ], 00:28:31.449 "driver_specific": { 00:28:31.449 "raid": { 00:28:31.449 "uuid": "0a49480c-abe0-4d7c-8b72-d1e96fa180d9", 00:28:31.449 "strip_size_kb": 0, 00:28:31.449 "state": "online", 00:28:31.449 "raid_level": "raid1", 00:28:31.449 "superblock": false, 00:28:31.449 "num_base_bdevs": 3, 00:28:31.449 "num_base_bdevs_discovered": 3, 00:28:31.449 "num_base_bdevs_operational": 3, 00:28:31.449 "base_bdevs_list": [ 00:28:31.449 { 00:28:31.449 "name": "NewBaseBdev", 00:28:31.449 "uuid": "b78aca2c-72ff-4759-b4f7-89fe90addb82", 00:28:31.449 "is_configured": true, 00:28:31.449 "data_offset": 0, 00:28:31.449 "data_size": 65536 00:28:31.449 }, 00:28:31.449 { 00:28:31.449 "name": "BaseBdev2", 00:28:31.449 "uuid": "875d72a0-bb69-4c8b-9edc-2c59ef06429a", 00:28:31.449 "is_configured": true, 00:28:31.449 "data_offset": 0, 00:28:31.449 "data_size": 65536 00:28:31.449 }, 00:28:31.449 { 00:28:31.449 "name": "BaseBdev3", 00:28:31.449 "uuid": "d827d362-f2f6-4fd5-8c35-2ac94076582a", 00:28:31.449 "is_configured": true, 00:28:31.449 "data_offset": 0, 00:28:31.449 "data_size": 65536 00:28:31.449 } 00:28:31.449 ] 00:28:31.449 } 00:28:31.449 } 00:28:31.449 }' 00:28:31.449 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:28:31.449 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:28:31.449 BaseBdev2 00:28:31.449 BaseBdev3' 00:28:31.449 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:31.707 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:28:31.707 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:31.707 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:31.707 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:31.708 [2024-11-20 13:48:38.603362] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:28:31.708 [2024-11-20 13:48:38.603402] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:28:31.708 [2024-11-20 13:48:38.603499] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:31.708 [2024-11-20 13:48:38.603760] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:31.708 [2024-11-20 13:48:38.603769] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 65732 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 65732 ']' 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 65732 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 65732 00:28:31.708 killing process with pid 65732 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 65732' 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 65732 00:28:31.708 [2024-11-20 13:48:38.632019] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:28:31.708 13:48:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 65732 00:28:31.966 [2024-11-20 13:48:38.787315] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:28:32.530 00:28:32.530 real 0m7.708s 00:28:32.530 user 0m12.344s 00:28:32.530 sys 0m1.263s 00:28:32.530 ************************************ 00:28:32.530 END TEST raid_state_function_test 00:28:32.530 ************************************ 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:32.530 13:48:39 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 3 true 00:28:32.530 13:48:39 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:28:32.530 13:48:39 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:28:32.530 13:48:39 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:28:32.530 ************************************ 00:28:32.530 START TEST raid_state_function_test_sb 00:28:32.530 ************************************ 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 3 true 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:28:32.530 Process raid pid: 66326 00:28:32.530 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=66326 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 66326' 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 66326 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 66326 ']' 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:28:32.530 13:48:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:32.530 [2024-11-20 13:48:39.520812] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:28:32.530 [2024-11-20 13:48:39.520934] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:28:32.787 [2024-11-20 13:48:39.677785] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:32.787 [2024-11-20 13:48:39.780374] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:28:33.043 [2024-11-20 13:48:39.903593] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:33.043 [2024-11-20 13:48:39.903642] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:33.607 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:28:33.607 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:28:33.607 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:28:33.607 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:33.607 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:33.607 [2024-11-20 13:48:40.375808] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:28:33.607 [2024-11-20 13:48:40.375871] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:28:33.607 [2024-11-20 13:48:40.375880] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:28:33.607 [2024-11-20 13:48:40.375889] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:28:33.607 [2024-11-20 13:48:40.375894] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:28:33.607 [2024-11-20 13:48:40.375902] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:28:33.607 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:33.607 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:28:33.607 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:33.607 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:33.607 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:33.607 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:33.607 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:33.607 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:33.607 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:33.607 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:33.607 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:33.607 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:33.607 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:33.607 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:33.607 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:33.607 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:33.607 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:33.607 "name": "Existed_Raid", 00:28:33.607 "uuid": "4a76c42f-4000-4e30-8f7a-e308b290b279", 00:28:33.607 "strip_size_kb": 0, 00:28:33.607 "state": "configuring", 00:28:33.607 "raid_level": "raid1", 00:28:33.607 "superblock": true, 00:28:33.607 "num_base_bdevs": 3, 00:28:33.607 "num_base_bdevs_discovered": 0, 00:28:33.607 "num_base_bdevs_operational": 3, 00:28:33.607 "base_bdevs_list": [ 00:28:33.607 { 00:28:33.607 "name": "BaseBdev1", 00:28:33.607 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:33.607 "is_configured": false, 00:28:33.607 "data_offset": 0, 00:28:33.607 "data_size": 0 00:28:33.607 }, 00:28:33.607 { 00:28:33.607 "name": "BaseBdev2", 00:28:33.607 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:33.607 "is_configured": false, 00:28:33.607 "data_offset": 0, 00:28:33.607 "data_size": 0 00:28:33.607 }, 00:28:33.607 { 00:28:33.607 "name": "BaseBdev3", 00:28:33.607 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:33.607 "is_configured": false, 00:28:33.607 "data_offset": 0, 00:28:33.607 "data_size": 0 00:28:33.607 } 00:28:33.607 ] 00:28:33.607 }' 00:28:33.607 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:33.607 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:33.931 [2024-11-20 13:48:40.695824] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:28:33.931 [2024-11-20 13:48:40.695867] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:33.931 [2024-11-20 13:48:40.703828] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:28:33.931 [2024-11-20 13:48:40.703876] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:28:33.931 [2024-11-20 13:48:40.703884] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:28:33.931 [2024-11-20 13:48:40.703892] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:28:33.931 [2024-11-20 13:48:40.703898] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:28:33.931 [2024-11-20 13:48:40.703905] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:33.931 [2024-11-20 13:48:40.734420] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:33.931 BaseBdev1 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:33.931 [ 00:28:33.931 { 00:28:33.931 "name": "BaseBdev1", 00:28:33.931 "aliases": [ 00:28:33.931 "1f1e4b53-f8ef-459e-a021-075cc045981f" 00:28:33.931 ], 00:28:33.931 "product_name": "Malloc disk", 00:28:33.931 "block_size": 512, 00:28:33.931 "num_blocks": 65536, 00:28:33.931 "uuid": "1f1e4b53-f8ef-459e-a021-075cc045981f", 00:28:33.931 "assigned_rate_limits": { 00:28:33.931 "rw_ios_per_sec": 0, 00:28:33.931 "rw_mbytes_per_sec": 0, 00:28:33.931 "r_mbytes_per_sec": 0, 00:28:33.931 "w_mbytes_per_sec": 0 00:28:33.931 }, 00:28:33.931 "claimed": true, 00:28:33.931 "claim_type": "exclusive_write", 00:28:33.931 "zoned": false, 00:28:33.931 "supported_io_types": { 00:28:33.931 "read": true, 00:28:33.931 "write": true, 00:28:33.931 "unmap": true, 00:28:33.931 "flush": true, 00:28:33.931 "reset": true, 00:28:33.931 "nvme_admin": false, 00:28:33.931 "nvme_io": false, 00:28:33.931 "nvme_io_md": false, 00:28:33.931 "write_zeroes": true, 00:28:33.931 "zcopy": true, 00:28:33.931 "get_zone_info": false, 00:28:33.931 "zone_management": false, 00:28:33.931 "zone_append": false, 00:28:33.931 "compare": false, 00:28:33.931 "compare_and_write": false, 00:28:33.931 "abort": true, 00:28:33.931 "seek_hole": false, 00:28:33.931 "seek_data": false, 00:28:33.931 "copy": true, 00:28:33.931 "nvme_iov_md": false 00:28:33.931 }, 00:28:33.931 "memory_domains": [ 00:28:33.931 { 00:28:33.931 "dma_device_id": "system", 00:28:33.931 "dma_device_type": 1 00:28:33.931 }, 00:28:33.931 { 00:28:33.931 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:33.931 "dma_device_type": 2 00:28:33.931 } 00:28:33.931 ], 00:28:33.931 "driver_specific": {} 00:28:33.931 } 00:28:33.931 ] 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:33.931 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:33.931 "name": "Existed_Raid", 00:28:33.931 "uuid": "46072808-607b-4002-91a2-ec4e9c7d3c03", 00:28:33.932 "strip_size_kb": 0, 00:28:33.932 "state": "configuring", 00:28:33.932 "raid_level": "raid1", 00:28:33.932 "superblock": true, 00:28:33.932 "num_base_bdevs": 3, 00:28:33.932 "num_base_bdevs_discovered": 1, 00:28:33.932 "num_base_bdevs_operational": 3, 00:28:33.932 "base_bdevs_list": [ 00:28:33.932 { 00:28:33.932 "name": "BaseBdev1", 00:28:33.932 "uuid": "1f1e4b53-f8ef-459e-a021-075cc045981f", 00:28:33.932 "is_configured": true, 00:28:33.932 "data_offset": 2048, 00:28:33.932 "data_size": 63488 00:28:33.932 }, 00:28:33.932 { 00:28:33.932 "name": "BaseBdev2", 00:28:33.932 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:33.932 "is_configured": false, 00:28:33.932 "data_offset": 0, 00:28:33.932 "data_size": 0 00:28:33.932 }, 00:28:33.932 { 00:28:33.932 "name": "BaseBdev3", 00:28:33.932 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:33.932 "is_configured": false, 00:28:33.932 "data_offset": 0, 00:28:33.932 "data_size": 0 00:28:33.932 } 00:28:33.932 ] 00:28:33.932 }' 00:28:33.932 13:48:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:33.932 13:48:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:34.204 [2024-11-20 13:48:41.090542] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:28:34.204 [2024-11-20 13:48:41.090599] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:34.204 [2024-11-20 13:48:41.102612] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:34.204 [2024-11-20 13:48:41.104366] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:28:34.204 [2024-11-20 13:48:41.104412] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:28:34.204 [2024-11-20 13:48:41.104420] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:28:34.204 [2024-11-20 13:48:41.104428] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:34.204 "name": "Existed_Raid", 00:28:34.204 "uuid": "80301e1c-e142-4fe9-b0e4-dca2be7b71e1", 00:28:34.204 "strip_size_kb": 0, 00:28:34.204 "state": "configuring", 00:28:34.204 "raid_level": "raid1", 00:28:34.204 "superblock": true, 00:28:34.204 "num_base_bdevs": 3, 00:28:34.204 "num_base_bdevs_discovered": 1, 00:28:34.204 "num_base_bdevs_operational": 3, 00:28:34.204 "base_bdevs_list": [ 00:28:34.204 { 00:28:34.204 "name": "BaseBdev1", 00:28:34.204 "uuid": "1f1e4b53-f8ef-459e-a021-075cc045981f", 00:28:34.204 "is_configured": true, 00:28:34.204 "data_offset": 2048, 00:28:34.204 "data_size": 63488 00:28:34.204 }, 00:28:34.204 { 00:28:34.204 "name": "BaseBdev2", 00:28:34.204 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:34.204 "is_configured": false, 00:28:34.204 "data_offset": 0, 00:28:34.204 "data_size": 0 00:28:34.204 }, 00:28:34.204 { 00:28:34.204 "name": "BaseBdev3", 00:28:34.204 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:34.204 "is_configured": false, 00:28:34.204 "data_offset": 0, 00:28:34.204 "data_size": 0 00:28:34.204 } 00:28:34.204 ] 00:28:34.204 }' 00:28:34.204 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:34.205 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:34.462 [2024-11-20 13:48:41.455409] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:28:34.462 BaseBdev2 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:34.462 [ 00:28:34.462 { 00:28:34.462 "name": "BaseBdev2", 00:28:34.462 "aliases": [ 00:28:34.462 "d61c8b62-3f6c-4641-ad98-ee273b1a4753" 00:28:34.462 ], 00:28:34.462 "product_name": "Malloc disk", 00:28:34.462 "block_size": 512, 00:28:34.462 "num_blocks": 65536, 00:28:34.462 "uuid": "d61c8b62-3f6c-4641-ad98-ee273b1a4753", 00:28:34.462 "assigned_rate_limits": { 00:28:34.462 "rw_ios_per_sec": 0, 00:28:34.462 "rw_mbytes_per_sec": 0, 00:28:34.462 "r_mbytes_per_sec": 0, 00:28:34.462 "w_mbytes_per_sec": 0 00:28:34.462 }, 00:28:34.462 "claimed": true, 00:28:34.462 "claim_type": "exclusive_write", 00:28:34.462 "zoned": false, 00:28:34.462 "supported_io_types": { 00:28:34.462 "read": true, 00:28:34.462 "write": true, 00:28:34.462 "unmap": true, 00:28:34.462 "flush": true, 00:28:34.462 "reset": true, 00:28:34.462 "nvme_admin": false, 00:28:34.462 "nvme_io": false, 00:28:34.462 "nvme_io_md": false, 00:28:34.462 "write_zeroes": true, 00:28:34.462 "zcopy": true, 00:28:34.462 "get_zone_info": false, 00:28:34.462 "zone_management": false, 00:28:34.462 "zone_append": false, 00:28:34.462 "compare": false, 00:28:34.462 "compare_and_write": false, 00:28:34.462 "abort": true, 00:28:34.462 "seek_hole": false, 00:28:34.462 "seek_data": false, 00:28:34.462 "copy": true, 00:28:34.462 "nvme_iov_md": false 00:28:34.462 }, 00:28:34.462 "memory_domains": [ 00:28:34.462 { 00:28:34.462 "dma_device_id": "system", 00:28:34.462 "dma_device_type": 1 00:28:34.462 }, 00:28:34.462 { 00:28:34.462 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:34.462 "dma_device_type": 2 00:28:34.462 } 00:28:34.462 ], 00:28:34.462 "driver_specific": {} 00:28:34.462 } 00:28:34.462 ] 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:34.462 "name": "Existed_Raid", 00:28:34.462 "uuid": "80301e1c-e142-4fe9-b0e4-dca2be7b71e1", 00:28:34.462 "strip_size_kb": 0, 00:28:34.462 "state": "configuring", 00:28:34.462 "raid_level": "raid1", 00:28:34.462 "superblock": true, 00:28:34.462 "num_base_bdevs": 3, 00:28:34.462 "num_base_bdevs_discovered": 2, 00:28:34.462 "num_base_bdevs_operational": 3, 00:28:34.462 "base_bdevs_list": [ 00:28:34.462 { 00:28:34.462 "name": "BaseBdev1", 00:28:34.462 "uuid": "1f1e4b53-f8ef-459e-a021-075cc045981f", 00:28:34.462 "is_configured": true, 00:28:34.462 "data_offset": 2048, 00:28:34.462 "data_size": 63488 00:28:34.462 }, 00:28:34.462 { 00:28:34.462 "name": "BaseBdev2", 00:28:34.462 "uuid": "d61c8b62-3f6c-4641-ad98-ee273b1a4753", 00:28:34.462 "is_configured": true, 00:28:34.462 "data_offset": 2048, 00:28:34.462 "data_size": 63488 00:28:34.462 }, 00:28:34.462 { 00:28:34.462 "name": "BaseBdev3", 00:28:34.462 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:34.462 "is_configured": false, 00:28:34.462 "data_offset": 0, 00:28:34.462 "data_size": 0 00:28:34.462 } 00:28:34.462 ] 00:28:34.462 }' 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:34.462 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:35.028 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:28:35.028 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:35.028 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:35.028 [2024-11-20 13:48:41.843879] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:28:35.028 [2024-11-20 13:48:41.844155] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:28:35.028 BaseBdev3 00:28:35.028 [2024-11-20 13:48:41.844176] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:28:35.028 [2024-11-20 13:48:41.844421] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:28:35.028 [2024-11-20 13:48:41.844553] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:28:35.028 [2024-11-20 13:48:41.844561] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:28:35.028 [2024-11-20 13:48:41.844684] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:35.028 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:35.028 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:28:35.028 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:28:35.028 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:35.028 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:28:35.028 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:35.028 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:35.028 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:35.028 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:35.028 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:35.028 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:35.028 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:28:35.028 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:35.028 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:35.028 [ 00:28:35.028 { 00:28:35.028 "name": "BaseBdev3", 00:28:35.028 "aliases": [ 00:28:35.028 "67a5485c-21cd-4ee2-81ba-f38c3061af64" 00:28:35.028 ], 00:28:35.028 "product_name": "Malloc disk", 00:28:35.028 "block_size": 512, 00:28:35.028 "num_blocks": 65536, 00:28:35.028 "uuid": "67a5485c-21cd-4ee2-81ba-f38c3061af64", 00:28:35.028 "assigned_rate_limits": { 00:28:35.028 "rw_ios_per_sec": 0, 00:28:35.028 "rw_mbytes_per_sec": 0, 00:28:35.028 "r_mbytes_per_sec": 0, 00:28:35.028 "w_mbytes_per_sec": 0 00:28:35.028 }, 00:28:35.028 "claimed": true, 00:28:35.028 "claim_type": "exclusive_write", 00:28:35.028 "zoned": false, 00:28:35.028 "supported_io_types": { 00:28:35.028 "read": true, 00:28:35.028 "write": true, 00:28:35.028 "unmap": true, 00:28:35.028 "flush": true, 00:28:35.028 "reset": true, 00:28:35.028 "nvme_admin": false, 00:28:35.028 "nvme_io": false, 00:28:35.028 "nvme_io_md": false, 00:28:35.028 "write_zeroes": true, 00:28:35.028 "zcopy": true, 00:28:35.028 "get_zone_info": false, 00:28:35.028 "zone_management": false, 00:28:35.028 "zone_append": false, 00:28:35.028 "compare": false, 00:28:35.028 "compare_and_write": false, 00:28:35.028 "abort": true, 00:28:35.028 "seek_hole": false, 00:28:35.028 "seek_data": false, 00:28:35.028 "copy": true, 00:28:35.028 "nvme_iov_md": false 00:28:35.028 }, 00:28:35.028 "memory_domains": [ 00:28:35.028 { 00:28:35.028 "dma_device_id": "system", 00:28:35.028 "dma_device_type": 1 00:28:35.028 }, 00:28:35.028 { 00:28:35.028 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:35.028 "dma_device_type": 2 00:28:35.028 } 00:28:35.028 ], 00:28:35.028 "driver_specific": {} 00:28:35.028 } 00:28:35.028 ] 00:28:35.028 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:35.028 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:28:35.028 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:28:35.028 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:28:35.028 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:28:35.028 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:35.028 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:35.028 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:35.028 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:35.029 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:35.029 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:35.029 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:35.029 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:35.029 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:35.029 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:35.029 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:35.029 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:35.029 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:35.029 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:35.029 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:35.029 "name": "Existed_Raid", 00:28:35.029 "uuid": "80301e1c-e142-4fe9-b0e4-dca2be7b71e1", 00:28:35.029 "strip_size_kb": 0, 00:28:35.029 "state": "online", 00:28:35.029 "raid_level": "raid1", 00:28:35.029 "superblock": true, 00:28:35.029 "num_base_bdevs": 3, 00:28:35.029 "num_base_bdevs_discovered": 3, 00:28:35.029 "num_base_bdevs_operational": 3, 00:28:35.029 "base_bdevs_list": [ 00:28:35.029 { 00:28:35.029 "name": "BaseBdev1", 00:28:35.029 "uuid": "1f1e4b53-f8ef-459e-a021-075cc045981f", 00:28:35.029 "is_configured": true, 00:28:35.029 "data_offset": 2048, 00:28:35.029 "data_size": 63488 00:28:35.029 }, 00:28:35.029 { 00:28:35.029 "name": "BaseBdev2", 00:28:35.029 "uuid": "d61c8b62-3f6c-4641-ad98-ee273b1a4753", 00:28:35.029 "is_configured": true, 00:28:35.029 "data_offset": 2048, 00:28:35.029 "data_size": 63488 00:28:35.029 }, 00:28:35.029 { 00:28:35.029 "name": "BaseBdev3", 00:28:35.029 "uuid": "67a5485c-21cd-4ee2-81ba-f38c3061af64", 00:28:35.029 "is_configured": true, 00:28:35.029 "data_offset": 2048, 00:28:35.029 "data_size": 63488 00:28:35.029 } 00:28:35.029 ] 00:28:35.029 }' 00:28:35.029 13:48:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:35.029 13:48:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:35.287 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:28:35.287 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:28:35.287 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:28:35.287 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:28:35.287 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:28:35.287 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:28:35.287 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:28:35.287 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:35.287 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:35.287 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:28:35.287 [2024-11-20 13:48:42.224318] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:35.287 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:35.287 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:35.287 "name": "Existed_Raid", 00:28:35.287 "aliases": [ 00:28:35.287 "80301e1c-e142-4fe9-b0e4-dca2be7b71e1" 00:28:35.287 ], 00:28:35.287 "product_name": "Raid Volume", 00:28:35.287 "block_size": 512, 00:28:35.287 "num_blocks": 63488, 00:28:35.287 "uuid": "80301e1c-e142-4fe9-b0e4-dca2be7b71e1", 00:28:35.287 "assigned_rate_limits": { 00:28:35.287 "rw_ios_per_sec": 0, 00:28:35.287 "rw_mbytes_per_sec": 0, 00:28:35.287 "r_mbytes_per_sec": 0, 00:28:35.287 "w_mbytes_per_sec": 0 00:28:35.287 }, 00:28:35.287 "claimed": false, 00:28:35.287 "zoned": false, 00:28:35.287 "supported_io_types": { 00:28:35.287 "read": true, 00:28:35.287 "write": true, 00:28:35.287 "unmap": false, 00:28:35.287 "flush": false, 00:28:35.287 "reset": true, 00:28:35.287 "nvme_admin": false, 00:28:35.287 "nvme_io": false, 00:28:35.287 "nvme_io_md": false, 00:28:35.287 "write_zeroes": true, 00:28:35.287 "zcopy": false, 00:28:35.287 "get_zone_info": false, 00:28:35.287 "zone_management": false, 00:28:35.287 "zone_append": false, 00:28:35.287 "compare": false, 00:28:35.287 "compare_and_write": false, 00:28:35.287 "abort": false, 00:28:35.287 "seek_hole": false, 00:28:35.287 "seek_data": false, 00:28:35.287 "copy": false, 00:28:35.287 "nvme_iov_md": false 00:28:35.287 }, 00:28:35.287 "memory_domains": [ 00:28:35.287 { 00:28:35.287 "dma_device_id": "system", 00:28:35.287 "dma_device_type": 1 00:28:35.287 }, 00:28:35.287 { 00:28:35.287 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:35.287 "dma_device_type": 2 00:28:35.287 }, 00:28:35.287 { 00:28:35.287 "dma_device_id": "system", 00:28:35.287 "dma_device_type": 1 00:28:35.287 }, 00:28:35.287 { 00:28:35.287 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:35.287 "dma_device_type": 2 00:28:35.287 }, 00:28:35.287 { 00:28:35.287 "dma_device_id": "system", 00:28:35.287 "dma_device_type": 1 00:28:35.287 }, 00:28:35.287 { 00:28:35.287 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:35.287 "dma_device_type": 2 00:28:35.287 } 00:28:35.287 ], 00:28:35.287 "driver_specific": { 00:28:35.287 "raid": { 00:28:35.287 "uuid": "80301e1c-e142-4fe9-b0e4-dca2be7b71e1", 00:28:35.287 "strip_size_kb": 0, 00:28:35.287 "state": "online", 00:28:35.287 "raid_level": "raid1", 00:28:35.287 "superblock": true, 00:28:35.287 "num_base_bdevs": 3, 00:28:35.287 "num_base_bdevs_discovered": 3, 00:28:35.287 "num_base_bdevs_operational": 3, 00:28:35.287 "base_bdevs_list": [ 00:28:35.287 { 00:28:35.287 "name": "BaseBdev1", 00:28:35.287 "uuid": "1f1e4b53-f8ef-459e-a021-075cc045981f", 00:28:35.287 "is_configured": true, 00:28:35.287 "data_offset": 2048, 00:28:35.287 "data_size": 63488 00:28:35.287 }, 00:28:35.287 { 00:28:35.287 "name": "BaseBdev2", 00:28:35.287 "uuid": "d61c8b62-3f6c-4641-ad98-ee273b1a4753", 00:28:35.287 "is_configured": true, 00:28:35.287 "data_offset": 2048, 00:28:35.287 "data_size": 63488 00:28:35.287 }, 00:28:35.287 { 00:28:35.287 "name": "BaseBdev3", 00:28:35.287 "uuid": "67a5485c-21cd-4ee2-81ba-f38c3061af64", 00:28:35.287 "is_configured": true, 00:28:35.287 "data_offset": 2048, 00:28:35.287 "data_size": 63488 00:28:35.287 } 00:28:35.288 ] 00:28:35.288 } 00:28:35.288 } 00:28:35.288 }' 00:28:35.288 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:28:35.288 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:28:35.288 BaseBdev2 00:28:35.288 BaseBdev3' 00:28:35.288 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:35.288 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:28:35.288 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:35.288 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:35.288 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:28:35.288 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:35.288 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:35.288 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:35.546 [2024-11-20 13:48:42.440137] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:35.546 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:35.546 "name": "Existed_Raid", 00:28:35.546 "uuid": "80301e1c-e142-4fe9-b0e4-dca2be7b71e1", 00:28:35.546 "strip_size_kb": 0, 00:28:35.546 "state": "online", 00:28:35.546 "raid_level": "raid1", 00:28:35.546 "superblock": true, 00:28:35.546 "num_base_bdevs": 3, 00:28:35.546 "num_base_bdevs_discovered": 2, 00:28:35.546 "num_base_bdevs_operational": 2, 00:28:35.546 "base_bdevs_list": [ 00:28:35.546 { 00:28:35.546 "name": null, 00:28:35.546 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:35.546 "is_configured": false, 00:28:35.546 "data_offset": 0, 00:28:35.546 "data_size": 63488 00:28:35.546 }, 00:28:35.546 { 00:28:35.546 "name": "BaseBdev2", 00:28:35.547 "uuid": "d61c8b62-3f6c-4641-ad98-ee273b1a4753", 00:28:35.547 "is_configured": true, 00:28:35.547 "data_offset": 2048, 00:28:35.547 "data_size": 63488 00:28:35.547 }, 00:28:35.547 { 00:28:35.547 "name": "BaseBdev3", 00:28:35.547 "uuid": "67a5485c-21cd-4ee2-81ba-f38c3061af64", 00:28:35.547 "is_configured": true, 00:28:35.547 "data_offset": 2048, 00:28:35.547 "data_size": 63488 00:28:35.547 } 00:28:35.547 ] 00:28:35.547 }' 00:28:35.547 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:35.547 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:35.803 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:28:35.803 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:28:35.803 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:35.803 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:28:35.803 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:35.803 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:35.804 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:35.804 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:28:35.804 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:28:35.804 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:28:35.804 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:35.804 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:35.804 [2024-11-20 13:48:42.850873] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:28:36.061 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.061 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:28:36.061 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:28:36.061 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:36.061 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:28:36.061 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.061 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:36.061 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.061 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:28:36.061 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:28:36.061 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:28:36.061 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.061 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:36.062 [2024-11-20 13:48:42.941432] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:28:36.062 [2024-11-20 13:48:42.941548] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:28:36.062 [2024-11-20 13:48:42.991918] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:36.062 [2024-11-20 13:48:42.992146] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:36.062 [2024-11-20 13:48:42.992217] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:28:36.062 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.062 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:28:36.062 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:28:36.062 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:28:36.062 13:48:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:36.062 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.062 13:48:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:36.062 BaseBdev2 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:36.062 [ 00:28:36.062 { 00:28:36.062 "name": "BaseBdev2", 00:28:36.062 "aliases": [ 00:28:36.062 "36f2416a-a9d5-48ad-98d4-76dea5054797" 00:28:36.062 ], 00:28:36.062 "product_name": "Malloc disk", 00:28:36.062 "block_size": 512, 00:28:36.062 "num_blocks": 65536, 00:28:36.062 "uuid": "36f2416a-a9d5-48ad-98d4-76dea5054797", 00:28:36.062 "assigned_rate_limits": { 00:28:36.062 "rw_ios_per_sec": 0, 00:28:36.062 "rw_mbytes_per_sec": 0, 00:28:36.062 "r_mbytes_per_sec": 0, 00:28:36.062 "w_mbytes_per_sec": 0 00:28:36.062 }, 00:28:36.062 "claimed": false, 00:28:36.062 "zoned": false, 00:28:36.062 "supported_io_types": { 00:28:36.062 "read": true, 00:28:36.062 "write": true, 00:28:36.062 "unmap": true, 00:28:36.062 "flush": true, 00:28:36.062 "reset": true, 00:28:36.062 "nvme_admin": false, 00:28:36.062 "nvme_io": false, 00:28:36.062 "nvme_io_md": false, 00:28:36.062 "write_zeroes": true, 00:28:36.062 "zcopy": true, 00:28:36.062 "get_zone_info": false, 00:28:36.062 "zone_management": false, 00:28:36.062 "zone_append": false, 00:28:36.062 "compare": false, 00:28:36.062 "compare_and_write": false, 00:28:36.062 "abort": true, 00:28:36.062 "seek_hole": false, 00:28:36.062 "seek_data": false, 00:28:36.062 "copy": true, 00:28:36.062 "nvme_iov_md": false 00:28:36.062 }, 00:28:36.062 "memory_domains": [ 00:28:36.062 { 00:28:36.062 "dma_device_id": "system", 00:28:36.062 "dma_device_type": 1 00:28:36.062 }, 00:28:36.062 { 00:28:36.062 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:36.062 "dma_device_type": 2 00:28:36.062 } 00:28:36.062 ], 00:28:36.062 "driver_specific": {} 00:28:36.062 } 00:28:36.062 ] 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.062 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:36.320 BaseBdev3 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:36.320 [ 00:28:36.320 { 00:28:36.320 "name": "BaseBdev3", 00:28:36.320 "aliases": [ 00:28:36.320 "4804dcea-6a08-44af-939e-c8dfaf1fdcbb" 00:28:36.320 ], 00:28:36.320 "product_name": "Malloc disk", 00:28:36.320 "block_size": 512, 00:28:36.320 "num_blocks": 65536, 00:28:36.320 "uuid": "4804dcea-6a08-44af-939e-c8dfaf1fdcbb", 00:28:36.320 "assigned_rate_limits": { 00:28:36.320 "rw_ios_per_sec": 0, 00:28:36.320 "rw_mbytes_per_sec": 0, 00:28:36.320 "r_mbytes_per_sec": 0, 00:28:36.320 "w_mbytes_per_sec": 0 00:28:36.320 }, 00:28:36.320 "claimed": false, 00:28:36.320 "zoned": false, 00:28:36.320 "supported_io_types": { 00:28:36.320 "read": true, 00:28:36.320 "write": true, 00:28:36.320 "unmap": true, 00:28:36.320 "flush": true, 00:28:36.320 "reset": true, 00:28:36.320 "nvme_admin": false, 00:28:36.320 "nvme_io": false, 00:28:36.320 "nvme_io_md": false, 00:28:36.320 "write_zeroes": true, 00:28:36.320 "zcopy": true, 00:28:36.320 "get_zone_info": false, 00:28:36.320 "zone_management": false, 00:28:36.320 "zone_append": false, 00:28:36.320 "compare": false, 00:28:36.320 "compare_and_write": false, 00:28:36.320 "abort": true, 00:28:36.320 "seek_hole": false, 00:28:36.320 "seek_data": false, 00:28:36.320 "copy": true, 00:28:36.320 "nvme_iov_md": false 00:28:36.320 }, 00:28:36.320 "memory_domains": [ 00:28:36.320 { 00:28:36.320 "dma_device_id": "system", 00:28:36.320 "dma_device_type": 1 00:28:36.320 }, 00:28:36.320 { 00:28:36.320 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:36.320 "dma_device_type": 2 00:28:36.320 } 00:28:36.320 ], 00:28:36.320 "driver_specific": {} 00:28:36.320 } 00:28:36.320 ] 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:36.320 [2024-11-20 13:48:43.152262] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:28:36.320 [2024-11-20 13:48:43.152441] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:28:36.320 [2024-11-20 13:48:43.152513] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:28:36.320 [2024-11-20 13:48:43.154328] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:36.320 "name": "Existed_Raid", 00:28:36.320 "uuid": "1d201d07-f45a-49cb-beb3-e06de5ec357f", 00:28:36.320 "strip_size_kb": 0, 00:28:36.320 "state": "configuring", 00:28:36.320 "raid_level": "raid1", 00:28:36.320 "superblock": true, 00:28:36.320 "num_base_bdevs": 3, 00:28:36.320 "num_base_bdevs_discovered": 2, 00:28:36.320 "num_base_bdevs_operational": 3, 00:28:36.320 "base_bdevs_list": [ 00:28:36.320 { 00:28:36.320 "name": "BaseBdev1", 00:28:36.320 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:36.320 "is_configured": false, 00:28:36.320 "data_offset": 0, 00:28:36.320 "data_size": 0 00:28:36.320 }, 00:28:36.320 { 00:28:36.320 "name": "BaseBdev2", 00:28:36.320 "uuid": "36f2416a-a9d5-48ad-98d4-76dea5054797", 00:28:36.320 "is_configured": true, 00:28:36.320 "data_offset": 2048, 00:28:36.320 "data_size": 63488 00:28:36.320 }, 00:28:36.320 { 00:28:36.320 "name": "BaseBdev3", 00:28:36.320 "uuid": "4804dcea-6a08-44af-939e-c8dfaf1fdcbb", 00:28:36.320 "is_configured": true, 00:28:36.320 "data_offset": 2048, 00:28:36.320 "data_size": 63488 00:28:36.320 } 00:28:36.320 ] 00:28:36.320 }' 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:36.320 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:36.578 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:28:36.578 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.578 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:36.578 [2024-11-20 13:48:43.480340] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:28:36.578 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.578 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:28:36.578 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:36.579 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:36.579 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:36.579 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:36.579 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:36.579 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:36.579 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:36.579 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:36.579 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:36.579 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:36.579 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.579 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:36.579 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:36.579 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.579 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:36.579 "name": "Existed_Raid", 00:28:36.579 "uuid": "1d201d07-f45a-49cb-beb3-e06de5ec357f", 00:28:36.579 "strip_size_kb": 0, 00:28:36.579 "state": "configuring", 00:28:36.579 "raid_level": "raid1", 00:28:36.579 "superblock": true, 00:28:36.579 "num_base_bdevs": 3, 00:28:36.579 "num_base_bdevs_discovered": 1, 00:28:36.579 "num_base_bdevs_operational": 3, 00:28:36.579 "base_bdevs_list": [ 00:28:36.579 { 00:28:36.579 "name": "BaseBdev1", 00:28:36.579 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:36.579 "is_configured": false, 00:28:36.579 "data_offset": 0, 00:28:36.579 "data_size": 0 00:28:36.579 }, 00:28:36.579 { 00:28:36.579 "name": null, 00:28:36.579 "uuid": "36f2416a-a9d5-48ad-98d4-76dea5054797", 00:28:36.579 "is_configured": false, 00:28:36.579 "data_offset": 0, 00:28:36.579 "data_size": 63488 00:28:36.579 }, 00:28:36.579 { 00:28:36.579 "name": "BaseBdev3", 00:28:36.579 "uuid": "4804dcea-6a08-44af-939e-c8dfaf1fdcbb", 00:28:36.579 "is_configured": true, 00:28:36.579 "data_offset": 2048, 00:28:36.579 "data_size": 63488 00:28:36.579 } 00:28:36.579 ] 00:28:36.579 }' 00:28:36.579 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:36.579 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:36.836 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:36.836 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.836 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:36.836 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:28:36.836 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.836 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:28:36.836 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:28:36.836 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.836 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:36.836 [2024-11-20 13:48:43.844942] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:36.836 BaseBdev1 00:28:36.836 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.836 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:28:36.836 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:28:36.836 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:36.836 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:28:36.836 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:36.836 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:36.836 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:36.836 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.836 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:36.836 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.836 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:28:36.836 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.836 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:36.837 [ 00:28:36.837 { 00:28:36.837 "name": "BaseBdev1", 00:28:36.837 "aliases": [ 00:28:36.837 "678634a5-9644-413e-8b3d-9b8e42287de4" 00:28:36.837 ], 00:28:36.837 "product_name": "Malloc disk", 00:28:36.837 "block_size": 512, 00:28:36.837 "num_blocks": 65536, 00:28:36.837 "uuid": "678634a5-9644-413e-8b3d-9b8e42287de4", 00:28:36.837 "assigned_rate_limits": { 00:28:36.837 "rw_ios_per_sec": 0, 00:28:36.837 "rw_mbytes_per_sec": 0, 00:28:36.837 "r_mbytes_per_sec": 0, 00:28:36.837 "w_mbytes_per_sec": 0 00:28:36.837 }, 00:28:36.837 "claimed": true, 00:28:36.837 "claim_type": "exclusive_write", 00:28:36.837 "zoned": false, 00:28:36.837 "supported_io_types": { 00:28:36.837 "read": true, 00:28:36.837 "write": true, 00:28:36.837 "unmap": true, 00:28:36.837 "flush": true, 00:28:36.837 "reset": true, 00:28:36.837 "nvme_admin": false, 00:28:36.837 "nvme_io": false, 00:28:36.837 "nvme_io_md": false, 00:28:36.837 "write_zeroes": true, 00:28:36.837 "zcopy": true, 00:28:36.837 "get_zone_info": false, 00:28:36.837 "zone_management": false, 00:28:36.837 "zone_append": false, 00:28:36.837 "compare": false, 00:28:36.837 "compare_and_write": false, 00:28:36.837 "abort": true, 00:28:36.837 "seek_hole": false, 00:28:36.837 "seek_data": false, 00:28:36.837 "copy": true, 00:28:36.837 "nvme_iov_md": false 00:28:36.837 }, 00:28:36.837 "memory_domains": [ 00:28:36.837 { 00:28:36.837 "dma_device_id": "system", 00:28:36.837 "dma_device_type": 1 00:28:36.837 }, 00:28:36.837 { 00:28:36.837 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:36.837 "dma_device_type": 2 00:28:36.837 } 00:28:36.837 ], 00:28:36.837 "driver_specific": {} 00:28:36.837 } 00:28:36.837 ] 00:28:36.837 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:36.837 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:28:36.837 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:28:36.837 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:36.837 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:36.837 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:36.837 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:36.837 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:36.837 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:36.837 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:36.837 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:36.837 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:36.837 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:36.837 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:36.837 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:36.837 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:36.837 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:37.132 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:37.132 "name": "Existed_Raid", 00:28:37.132 "uuid": "1d201d07-f45a-49cb-beb3-e06de5ec357f", 00:28:37.132 "strip_size_kb": 0, 00:28:37.132 "state": "configuring", 00:28:37.133 "raid_level": "raid1", 00:28:37.133 "superblock": true, 00:28:37.133 "num_base_bdevs": 3, 00:28:37.133 "num_base_bdevs_discovered": 2, 00:28:37.133 "num_base_bdevs_operational": 3, 00:28:37.133 "base_bdevs_list": [ 00:28:37.133 { 00:28:37.133 "name": "BaseBdev1", 00:28:37.133 "uuid": "678634a5-9644-413e-8b3d-9b8e42287de4", 00:28:37.133 "is_configured": true, 00:28:37.133 "data_offset": 2048, 00:28:37.133 "data_size": 63488 00:28:37.133 }, 00:28:37.133 { 00:28:37.133 "name": null, 00:28:37.133 "uuid": "36f2416a-a9d5-48ad-98d4-76dea5054797", 00:28:37.133 "is_configured": false, 00:28:37.133 "data_offset": 0, 00:28:37.133 "data_size": 63488 00:28:37.133 }, 00:28:37.133 { 00:28:37.133 "name": "BaseBdev3", 00:28:37.133 "uuid": "4804dcea-6a08-44af-939e-c8dfaf1fdcbb", 00:28:37.133 "is_configured": true, 00:28:37.133 "data_offset": 2048, 00:28:37.133 "data_size": 63488 00:28:37.133 } 00:28:37.133 ] 00:28:37.133 }' 00:28:37.133 13:48:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:37.133 13:48:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:37.394 [2024-11-20 13:48:44.225080] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:37.394 "name": "Existed_Raid", 00:28:37.394 "uuid": "1d201d07-f45a-49cb-beb3-e06de5ec357f", 00:28:37.394 "strip_size_kb": 0, 00:28:37.394 "state": "configuring", 00:28:37.394 "raid_level": "raid1", 00:28:37.394 "superblock": true, 00:28:37.394 "num_base_bdevs": 3, 00:28:37.394 "num_base_bdevs_discovered": 1, 00:28:37.394 "num_base_bdevs_operational": 3, 00:28:37.394 "base_bdevs_list": [ 00:28:37.394 { 00:28:37.394 "name": "BaseBdev1", 00:28:37.394 "uuid": "678634a5-9644-413e-8b3d-9b8e42287de4", 00:28:37.394 "is_configured": true, 00:28:37.394 "data_offset": 2048, 00:28:37.394 "data_size": 63488 00:28:37.394 }, 00:28:37.394 { 00:28:37.394 "name": null, 00:28:37.394 "uuid": "36f2416a-a9d5-48ad-98d4-76dea5054797", 00:28:37.394 "is_configured": false, 00:28:37.394 "data_offset": 0, 00:28:37.394 "data_size": 63488 00:28:37.394 }, 00:28:37.394 { 00:28:37.394 "name": null, 00:28:37.394 "uuid": "4804dcea-6a08-44af-939e-c8dfaf1fdcbb", 00:28:37.394 "is_configured": false, 00:28:37.394 "data_offset": 0, 00:28:37.394 "data_size": 63488 00:28:37.394 } 00:28:37.394 ] 00:28:37.394 }' 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:37.394 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:37.653 [2024-11-20 13:48:44.573197] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:37.653 "name": "Existed_Raid", 00:28:37.653 "uuid": "1d201d07-f45a-49cb-beb3-e06de5ec357f", 00:28:37.653 "strip_size_kb": 0, 00:28:37.653 "state": "configuring", 00:28:37.653 "raid_level": "raid1", 00:28:37.653 "superblock": true, 00:28:37.653 "num_base_bdevs": 3, 00:28:37.653 "num_base_bdevs_discovered": 2, 00:28:37.653 "num_base_bdevs_operational": 3, 00:28:37.653 "base_bdevs_list": [ 00:28:37.653 { 00:28:37.653 "name": "BaseBdev1", 00:28:37.653 "uuid": "678634a5-9644-413e-8b3d-9b8e42287de4", 00:28:37.653 "is_configured": true, 00:28:37.653 "data_offset": 2048, 00:28:37.653 "data_size": 63488 00:28:37.653 }, 00:28:37.653 { 00:28:37.653 "name": null, 00:28:37.653 "uuid": "36f2416a-a9d5-48ad-98d4-76dea5054797", 00:28:37.653 "is_configured": false, 00:28:37.653 "data_offset": 0, 00:28:37.653 "data_size": 63488 00:28:37.653 }, 00:28:37.653 { 00:28:37.653 "name": "BaseBdev3", 00:28:37.653 "uuid": "4804dcea-6a08-44af-939e-c8dfaf1fdcbb", 00:28:37.653 "is_configured": true, 00:28:37.653 "data_offset": 2048, 00:28:37.653 "data_size": 63488 00:28:37.653 } 00:28:37.653 ] 00:28:37.653 }' 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:37.653 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:37.911 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:37.911 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:28:37.911 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:37.911 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:37.911 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:37.911 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:28:37.911 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:28:37.911 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:37.911 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:37.911 [2024-11-20 13:48:44.941272] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:28:38.170 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:38.170 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:28:38.170 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:38.170 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:38.170 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:38.170 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:38.170 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:38.170 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:38.170 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:38.170 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:38.170 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:38.170 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:38.170 13:48:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:38.170 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:38.170 13:48:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:38.170 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:38.170 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:38.170 "name": "Existed_Raid", 00:28:38.170 "uuid": "1d201d07-f45a-49cb-beb3-e06de5ec357f", 00:28:38.170 "strip_size_kb": 0, 00:28:38.170 "state": "configuring", 00:28:38.170 "raid_level": "raid1", 00:28:38.170 "superblock": true, 00:28:38.170 "num_base_bdevs": 3, 00:28:38.170 "num_base_bdevs_discovered": 1, 00:28:38.170 "num_base_bdevs_operational": 3, 00:28:38.170 "base_bdevs_list": [ 00:28:38.170 { 00:28:38.170 "name": null, 00:28:38.170 "uuid": "678634a5-9644-413e-8b3d-9b8e42287de4", 00:28:38.170 "is_configured": false, 00:28:38.170 "data_offset": 0, 00:28:38.170 "data_size": 63488 00:28:38.170 }, 00:28:38.170 { 00:28:38.170 "name": null, 00:28:38.170 "uuid": "36f2416a-a9d5-48ad-98d4-76dea5054797", 00:28:38.170 "is_configured": false, 00:28:38.170 "data_offset": 0, 00:28:38.170 "data_size": 63488 00:28:38.170 }, 00:28:38.170 { 00:28:38.170 "name": "BaseBdev3", 00:28:38.170 "uuid": "4804dcea-6a08-44af-939e-c8dfaf1fdcbb", 00:28:38.170 "is_configured": true, 00:28:38.170 "data_offset": 2048, 00:28:38.170 "data_size": 63488 00:28:38.170 } 00:28:38.170 ] 00:28:38.170 }' 00:28:38.170 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:38.170 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:38.429 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:28:38.429 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:38.429 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:38.429 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:38.429 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:38.429 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:28:38.429 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:28:38.429 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:38.429 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:38.429 [2024-11-20 13:48:45.334591] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:28:38.429 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:38.429 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:28:38.429 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:38.429 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:38.429 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:38.429 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:38.429 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:38.429 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:38.429 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:38.429 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:38.429 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:38.429 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:38.429 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:38.429 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:38.429 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:38.429 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:38.429 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:38.429 "name": "Existed_Raid", 00:28:38.429 "uuid": "1d201d07-f45a-49cb-beb3-e06de5ec357f", 00:28:38.429 "strip_size_kb": 0, 00:28:38.429 "state": "configuring", 00:28:38.430 "raid_level": "raid1", 00:28:38.430 "superblock": true, 00:28:38.430 "num_base_bdevs": 3, 00:28:38.430 "num_base_bdevs_discovered": 2, 00:28:38.430 "num_base_bdevs_operational": 3, 00:28:38.430 "base_bdevs_list": [ 00:28:38.430 { 00:28:38.430 "name": null, 00:28:38.430 "uuid": "678634a5-9644-413e-8b3d-9b8e42287de4", 00:28:38.430 "is_configured": false, 00:28:38.430 "data_offset": 0, 00:28:38.430 "data_size": 63488 00:28:38.430 }, 00:28:38.430 { 00:28:38.430 "name": "BaseBdev2", 00:28:38.430 "uuid": "36f2416a-a9d5-48ad-98d4-76dea5054797", 00:28:38.430 "is_configured": true, 00:28:38.430 "data_offset": 2048, 00:28:38.430 "data_size": 63488 00:28:38.430 }, 00:28:38.430 { 00:28:38.430 "name": "BaseBdev3", 00:28:38.430 "uuid": "4804dcea-6a08-44af-939e-c8dfaf1fdcbb", 00:28:38.430 "is_configured": true, 00:28:38.430 "data_offset": 2048, 00:28:38.430 "data_size": 63488 00:28:38.430 } 00:28:38.430 ] 00:28:38.430 }' 00:28:38.430 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:38.430 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 678634a5-9644-413e-8b3d-9b8e42287de4 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:38.688 [2024-11-20 13:48:45.731155] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:28:38.688 [2024-11-20 13:48:45.731362] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:28:38.688 [2024-11-20 13:48:45.731373] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:28:38.688 [2024-11-20 13:48:45.731613] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:28:38.688 NewBaseBdev 00:28:38.688 [2024-11-20 13:48:45.731732] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:28:38.688 [2024-11-20 13:48:45.731742] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:28:38.688 [2024-11-20 13:48:45.731850] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:38.688 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:38.947 [ 00:28:38.947 { 00:28:38.947 "name": "NewBaseBdev", 00:28:38.947 "aliases": [ 00:28:38.947 "678634a5-9644-413e-8b3d-9b8e42287de4" 00:28:38.947 ], 00:28:38.947 "product_name": "Malloc disk", 00:28:38.947 "block_size": 512, 00:28:38.947 "num_blocks": 65536, 00:28:38.947 "uuid": "678634a5-9644-413e-8b3d-9b8e42287de4", 00:28:38.947 "assigned_rate_limits": { 00:28:38.947 "rw_ios_per_sec": 0, 00:28:38.947 "rw_mbytes_per_sec": 0, 00:28:38.947 "r_mbytes_per_sec": 0, 00:28:38.947 "w_mbytes_per_sec": 0 00:28:38.947 }, 00:28:38.947 "claimed": true, 00:28:38.947 "claim_type": "exclusive_write", 00:28:38.947 "zoned": false, 00:28:38.947 "supported_io_types": { 00:28:38.947 "read": true, 00:28:38.947 "write": true, 00:28:38.947 "unmap": true, 00:28:38.947 "flush": true, 00:28:38.947 "reset": true, 00:28:38.947 "nvme_admin": false, 00:28:38.947 "nvme_io": false, 00:28:38.947 "nvme_io_md": false, 00:28:38.947 "write_zeroes": true, 00:28:38.947 "zcopy": true, 00:28:38.947 "get_zone_info": false, 00:28:38.947 "zone_management": false, 00:28:38.947 "zone_append": false, 00:28:38.947 "compare": false, 00:28:38.947 "compare_and_write": false, 00:28:38.947 "abort": true, 00:28:38.947 "seek_hole": false, 00:28:38.947 "seek_data": false, 00:28:38.947 "copy": true, 00:28:38.947 "nvme_iov_md": false 00:28:38.947 }, 00:28:38.947 "memory_domains": [ 00:28:38.947 { 00:28:38.947 "dma_device_id": "system", 00:28:38.947 "dma_device_type": 1 00:28:38.947 }, 00:28:38.947 { 00:28:38.947 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:38.947 "dma_device_type": 2 00:28:38.947 } 00:28:38.947 ], 00:28:38.947 "driver_specific": {} 00:28:38.947 } 00:28:38.947 ] 00:28:38.947 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:38.947 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:28:38.947 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:28:38.947 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:38.947 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:38.947 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:38.947 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:38.947 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:38.947 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:38.947 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:38.947 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:38.947 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:38.947 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:38.947 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:38.947 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:38.947 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:38.947 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:38.947 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:38.947 "name": "Existed_Raid", 00:28:38.947 "uuid": "1d201d07-f45a-49cb-beb3-e06de5ec357f", 00:28:38.947 "strip_size_kb": 0, 00:28:38.947 "state": "online", 00:28:38.947 "raid_level": "raid1", 00:28:38.947 "superblock": true, 00:28:38.947 "num_base_bdevs": 3, 00:28:38.947 "num_base_bdevs_discovered": 3, 00:28:38.947 "num_base_bdevs_operational": 3, 00:28:38.947 "base_bdevs_list": [ 00:28:38.947 { 00:28:38.947 "name": "NewBaseBdev", 00:28:38.947 "uuid": "678634a5-9644-413e-8b3d-9b8e42287de4", 00:28:38.947 "is_configured": true, 00:28:38.947 "data_offset": 2048, 00:28:38.947 "data_size": 63488 00:28:38.948 }, 00:28:38.948 { 00:28:38.948 "name": "BaseBdev2", 00:28:38.948 "uuid": "36f2416a-a9d5-48ad-98d4-76dea5054797", 00:28:38.948 "is_configured": true, 00:28:38.948 "data_offset": 2048, 00:28:38.948 "data_size": 63488 00:28:38.948 }, 00:28:38.948 { 00:28:38.948 "name": "BaseBdev3", 00:28:38.948 "uuid": "4804dcea-6a08-44af-939e-c8dfaf1fdcbb", 00:28:38.948 "is_configured": true, 00:28:38.948 "data_offset": 2048, 00:28:38.948 "data_size": 63488 00:28:38.948 } 00:28:38.948 ] 00:28:38.948 }' 00:28:38.948 13:48:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:38.948 13:48:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:39.206 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:28:39.206 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:28:39.206 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:28:39.206 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:28:39.206 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:28:39.206 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:28:39.206 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:28:39.206 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:28:39.206 13:48:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:39.206 13:48:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:39.206 [2024-11-20 13:48:46.083579] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:39.206 13:48:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:39.206 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:39.206 "name": "Existed_Raid", 00:28:39.206 "aliases": [ 00:28:39.206 "1d201d07-f45a-49cb-beb3-e06de5ec357f" 00:28:39.206 ], 00:28:39.206 "product_name": "Raid Volume", 00:28:39.206 "block_size": 512, 00:28:39.206 "num_blocks": 63488, 00:28:39.206 "uuid": "1d201d07-f45a-49cb-beb3-e06de5ec357f", 00:28:39.206 "assigned_rate_limits": { 00:28:39.206 "rw_ios_per_sec": 0, 00:28:39.206 "rw_mbytes_per_sec": 0, 00:28:39.206 "r_mbytes_per_sec": 0, 00:28:39.206 "w_mbytes_per_sec": 0 00:28:39.206 }, 00:28:39.206 "claimed": false, 00:28:39.206 "zoned": false, 00:28:39.206 "supported_io_types": { 00:28:39.206 "read": true, 00:28:39.206 "write": true, 00:28:39.206 "unmap": false, 00:28:39.206 "flush": false, 00:28:39.206 "reset": true, 00:28:39.206 "nvme_admin": false, 00:28:39.206 "nvme_io": false, 00:28:39.206 "nvme_io_md": false, 00:28:39.206 "write_zeroes": true, 00:28:39.206 "zcopy": false, 00:28:39.206 "get_zone_info": false, 00:28:39.206 "zone_management": false, 00:28:39.206 "zone_append": false, 00:28:39.206 "compare": false, 00:28:39.206 "compare_and_write": false, 00:28:39.206 "abort": false, 00:28:39.206 "seek_hole": false, 00:28:39.206 "seek_data": false, 00:28:39.206 "copy": false, 00:28:39.206 "nvme_iov_md": false 00:28:39.206 }, 00:28:39.206 "memory_domains": [ 00:28:39.206 { 00:28:39.206 "dma_device_id": "system", 00:28:39.206 "dma_device_type": 1 00:28:39.206 }, 00:28:39.206 { 00:28:39.206 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:39.206 "dma_device_type": 2 00:28:39.206 }, 00:28:39.206 { 00:28:39.206 "dma_device_id": "system", 00:28:39.206 "dma_device_type": 1 00:28:39.206 }, 00:28:39.206 { 00:28:39.206 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:39.206 "dma_device_type": 2 00:28:39.206 }, 00:28:39.206 { 00:28:39.206 "dma_device_id": "system", 00:28:39.206 "dma_device_type": 1 00:28:39.206 }, 00:28:39.206 { 00:28:39.206 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:39.206 "dma_device_type": 2 00:28:39.206 } 00:28:39.206 ], 00:28:39.206 "driver_specific": { 00:28:39.206 "raid": { 00:28:39.206 "uuid": "1d201d07-f45a-49cb-beb3-e06de5ec357f", 00:28:39.206 "strip_size_kb": 0, 00:28:39.206 "state": "online", 00:28:39.206 "raid_level": "raid1", 00:28:39.206 "superblock": true, 00:28:39.206 "num_base_bdevs": 3, 00:28:39.206 "num_base_bdevs_discovered": 3, 00:28:39.206 "num_base_bdevs_operational": 3, 00:28:39.206 "base_bdevs_list": [ 00:28:39.206 { 00:28:39.206 "name": "NewBaseBdev", 00:28:39.206 "uuid": "678634a5-9644-413e-8b3d-9b8e42287de4", 00:28:39.206 "is_configured": true, 00:28:39.206 "data_offset": 2048, 00:28:39.206 "data_size": 63488 00:28:39.206 }, 00:28:39.206 { 00:28:39.206 "name": "BaseBdev2", 00:28:39.206 "uuid": "36f2416a-a9d5-48ad-98d4-76dea5054797", 00:28:39.206 "is_configured": true, 00:28:39.206 "data_offset": 2048, 00:28:39.206 "data_size": 63488 00:28:39.206 }, 00:28:39.206 { 00:28:39.206 "name": "BaseBdev3", 00:28:39.206 "uuid": "4804dcea-6a08-44af-939e-c8dfaf1fdcbb", 00:28:39.206 "is_configured": true, 00:28:39.207 "data_offset": 2048, 00:28:39.207 "data_size": 63488 00:28:39.207 } 00:28:39.207 ] 00:28:39.207 } 00:28:39.207 } 00:28:39.207 }' 00:28:39.207 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:28:39.207 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:28:39.207 BaseBdev2 00:28:39.207 BaseBdev3' 00:28:39.207 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:39.207 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:28:39.207 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:39.207 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:28:39.207 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:39.207 13:48:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:39.207 13:48:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:39.207 13:48:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:39.207 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:39.207 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:39.207 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:39.207 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:28:39.207 13:48:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:39.207 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:39.207 13:48:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:39.207 13:48:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:39.207 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:39.207 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:39.207 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:39.207 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:28:39.207 13:48:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:39.207 13:48:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:39.207 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:39.207 13:48:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:39.465 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:39.465 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:39.465 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:28:39.465 13:48:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:39.465 13:48:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:39.465 [2024-11-20 13:48:46.271316] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:28:39.465 [2024-11-20 13:48:46.271354] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:28:39.465 [2024-11-20 13:48:46.271427] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:39.465 [2024-11-20 13:48:46.271702] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:39.465 [2024-11-20 13:48:46.271712] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:28:39.465 13:48:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:39.465 13:48:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 66326 00:28:39.465 13:48:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 66326 ']' 00:28:39.465 13:48:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 66326 00:28:39.465 13:48:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:28:39.465 13:48:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:28:39.465 13:48:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 66326 00:28:39.465 killing process with pid 66326 00:28:39.465 13:48:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:28:39.465 13:48:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:28:39.465 13:48:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 66326' 00:28:39.465 13:48:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 66326 00:28:39.465 [2024-11-20 13:48:46.302779] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:28:39.465 13:48:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 66326 00:28:39.465 [2024-11-20 13:48:46.461775] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:28:40.398 13:48:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:28:40.398 00:28:40.398 real 0m7.642s 00:28:40.398 user 0m12.225s 00:28:40.398 sys 0m1.329s 00:28:40.398 ************************************ 00:28:40.398 END TEST raid_state_function_test_sb 00:28:40.398 ************************************ 00:28:40.398 13:48:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:28:40.398 13:48:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:40.398 13:48:47 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 3 00:28:40.398 13:48:47 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:28:40.398 13:48:47 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:28:40.398 13:48:47 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:28:40.398 ************************************ 00:28:40.398 START TEST raid_superblock_test 00:28:40.398 ************************************ 00:28:40.398 13:48:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 3 00:28:40.398 13:48:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:28:40.398 13:48:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:28:40.398 13:48:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:28:40.398 13:48:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:28:40.398 13:48:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:28:40.398 13:48:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:28:40.398 13:48:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:28:40.398 13:48:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:28:40.398 13:48:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:28:40.398 13:48:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:28:40.398 13:48:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:28:40.398 13:48:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:28:40.398 13:48:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:28:40.398 13:48:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:28:40.398 13:48:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:28:40.398 13:48:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=66919 00:28:40.398 13:48:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:28:40.398 13:48:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 66919 00:28:40.398 13:48:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 66919 ']' 00:28:40.398 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:28:40.398 13:48:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:28:40.398 13:48:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:28:40.398 13:48:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:28:40.398 13:48:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:28:40.398 13:48:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:40.398 [2024-11-20 13:48:47.231937] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:28:40.398 [2024-11-20 13:48:47.232125] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid66919 ] 00:28:40.398 [2024-11-20 13:48:47.401863] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:40.655 [2024-11-20 13:48:47.527042] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:28:40.655 [2024-11-20 13:48:47.675796] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:40.655 [2024-11-20 13:48:47.676058] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:41.221 malloc1 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:41.221 [2024-11-20 13:48:48.109003] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:28:41.221 [2024-11-20 13:48:48.109084] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:41.221 [2024-11-20 13:48:48.109109] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:28:41.221 [2024-11-20 13:48:48.109119] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:41.221 [2024-11-20 13:48:48.111595] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:41.221 [2024-11-20 13:48:48.111815] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:28:41.221 pt1 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:41.221 malloc2 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:41.221 [2024-11-20 13:48:48.156073] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:28:41.221 [2024-11-20 13:48:48.156153] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:41.221 [2024-11-20 13:48:48.156184] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:28:41.221 [2024-11-20 13:48:48.156195] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:41.221 [2024-11-20 13:48:48.158585] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:41.221 [2024-11-20 13:48:48.158624] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:28:41.221 pt2 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:41.221 malloc3 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:41.221 [2024-11-20 13:48:48.212107] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:28:41.221 [2024-11-20 13:48:48.212292] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:41.221 [2024-11-20 13:48:48.212327] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:28:41.221 [2024-11-20 13:48:48.212339] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:41.221 [2024-11-20 13:48:48.214781] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:41.221 [2024-11-20 13:48:48.214823] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:28:41.221 pt3 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:28:41.221 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:28:41.222 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:28:41.222 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:41.222 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:41.222 [2024-11-20 13:48:48.220097] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:28:41.222 [2024-11-20 13:48:48.222129] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:28:41.222 [2024-11-20 13:48:48.222197] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:28:41.222 [2024-11-20 13:48:48.222381] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:28:41.222 [2024-11-20 13:48:48.222399] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:28:41.222 [2024-11-20 13:48:48.222688] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:28:41.222 [2024-11-20 13:48:48.222854] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:28:41.222 [2024-11-20 13:48:48.222869] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:28:41.222 [2024-11-20 13:48:48.223076] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:41.222 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:41.222 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:28:41.222 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:41.222 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:41.222 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:41.222 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:41.222 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:41.222 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:41.222 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:41.222 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:41.222 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:41.222 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:41.222 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:41.222 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:41.222 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:41.222 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:41.222 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:41.222 "name": "raid_bdev1", 00:28:41.222 "uuid": "da5832c9-15ce-413d-9988-86373121ff09", 00:28:41.222 "strip_size_kb": 0, 00:28:41.222 "state": "online", 00:28:41.222 "raid_level": "raid1", 00:28:41.222 "superblock": true, 00:28:41.222 "num_base_bdevs": 3, 00:28:41.222 "num_base_bdevs_discovered": 3, 00:28:41.222 "num_base_bdevs_operational": 3, 00:28:41.222 "base_bdevs_list": [ 00:28:41.222 { 00:28:41.222 "name": "pt1", 00:28:41.222 "uuid": "00000000-0000-0000-0000-000000000001", 00:28:41.222 "is_configured": true, 00:28:41.222 "data_offset": 2048, 00:28:41.222 "data_size": 63488 00:28:41.222 }, 00:28:41.222 { 00:28:41.222 "name": "pt2", 00:28:41.222 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:41.222 "is_configured": true, 00:28:41.222 "data_offset": 2048, 00:28:41.222 "data_size": 63488 00:28:41.222 }, 00:28:41.222 { 00:28:41.222 "name": "pt3", 00:28:41.222 "uuid": "00000000-0000-0000-0000-000000000003", 00:28:41.222 "is_configured": true, 00:28:41.222 "data_offset": 2048, 00:28:41.222 "data_size": 63488 00:28:41.222 } 00:28:41.222 ] 00:28:41.222 }' 00:28:41.222 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:41.222 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:41.789 [2024-11-20 13:48:48.568495] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:41.789 "name": "raid_bdev1", 00:28:41.789 "aliases": [ 00:28:41.789 "da5832c9-15ce-413d-9988-86373121ff09" 00:28:41.789 ], 00:28:41.789 "product_name": "Raid Volume", 00:28:41.789 "block_size": 512, 00:28:41.789 "num_blocks": 63488, 00:28:41.789 "uuid": "da5832c9-15ce-413d-9988-86373121ff09", 00:28:41.789 "assigned_rate_limits": { 00:28:41.789 "rw_ios_per_sec": 0, 00:28:41.789 "rw_mbytes_per_sec": 0, 00:28:41.789 "r_mbytes_per_sec": 0, 00:28:41.789 "w_mbytes_per_sec": 0 00:28:41.789 }, 00:28:41.789 "claimed": false, 00:28:41.789 "zoned": false, 00:28:41.789 "supported_io_types": { 00:28:41.789 "read": true, 00:28:41.789 "write": true, 00:28:41.789 "unmap": false, 00:28:41.789 "flush": false, 00:28:41.789 "reset": true, 00:28:41.789 "nvme_admin": false, 00:28:41.789 "nvme_io": false, 00:28:41.789 "nvme_io_md": false, 00:28:41.789 "write_zeroes": true, 00:28:41.789 "zcopy": false, 00:28:41.789 "get_zone_info": false, 00:28:41.789 "zone_management": false, 00:28:41.789 "zone_append": false, 00:28:41.789 "compare": false, 00:28:41.789 "compare_and_write": false, 00:28:41.789 "abort": false, 00:28:41.789 "seek_hole": false, 00:28:41.789 "seek_data": false, 00:28:41.789 "copy": false, 00:28:41.789 "nvme_iov_md": false 00:28:41.789 }, 00:28:41.789 "memory_domains": [ 00:28:41.789 { 00:28:41.789 "dma_device_id": "system", 00:28:41.789 "dma_device_type": 1 00:28:41.789 }, 00:28:41.789 { 00:28:41.789 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:41.789 "dma_device_type": 2 00:28:41.789 }, 00:28:41.789 { 00:28:41.789 "dma_device_id": "system", 00:28:41.789 "dma_device_type": 1 00:28:41.789 }, 00:28:41.789 { 00:28:41.789 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:41.789 "dma_device_type": 2 00:28:41.789 }, 00:28:41.789 { 00:28:41.789 "dma_device_id": "system", 00:28:41.789 "dma_device_type": 1 00:28:41.789 }, 00:28:41.789 { 00:28:41.789 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:41.789 "dma_device_type": 2 00:28:41.789 } 00:28:41.789 ], 00:28:41.789 "driver_specific": { 00:28:41.789 "raid": { 00:28:41.789 "uuid": "da5832c9-15ce-413d-9988-86373121ff09", 00:28:41.789 "strip_size_kb": 0, 00:28:41.789 "state": "online", 00:28:41.789 "raid_level": "raid1", 00:28:41.789 "superblock": true, 00:28:41.789 "num_base_bdevs": 3, 00:28:41.789 "num_base_bdevs_discovered": 3, 00:28:41.789 "num_base_bdevs_operational": 3, 00:28:41.789 "base_bdevs_list": [ 00:28:41.789 { 00:28:41.789 "name": "pt1", 00:28:41.789 "uuid": "00000000-0000-0000-0000-000000000001", 00:28:41.789 "is_configured": true, 00:28:41.789 "data_offset": 2048, 00:28:41.789 "data_size": 63488 00:28:41.789 }, 00:28:41.789 { 00:28:41.789 "name": "pt2", 00:28:41.789 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:41.789 "is_configured": true, 00:28:41.789 "data_offset": 2048, 00:28:41.789 "data_size": 63488 00:28:41.789 }, 00:28:41.789 { 00:28:41.789 "name": "pt3", 00:28:41.789 "uuid": "00000000-0000-0000-0000-000000000003", 00:28:41.789 "is_configured": true, 00:28:41.789 "data_offset": 2048, 00:28:41.789 "data_size": 63488 00:28:41.789 } 00:28:41.789 ] 00:28:41.789 } 00:28:41.789 } 00:28:41.789 }' 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:28:41.789 pt2 00:28:41.789 pt3' 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:41.789 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:28:41.790 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:41.790 [2024-11-20 13:48:48.768496] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:41.790 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:41.790 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=da5832c9-15ce-413d-9988-86373121ff09 00:28:41.790 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z da5832c9-15ce-413d-9988-86373121ff09 ']' 00:28:41.790 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:28:41.790 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:41.790 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:41.790 [2024-11-20 13:48:48.800188] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:28:41.790 [2024-11-20 13:48:48.800373] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:28:41.790 [2024-11-20 13:48:48.800480] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:41.790 [2024-11-20 13:48:48.800578] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:41.790 [2024-11-20 13:48:48.800590] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:28:41.790 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:41.790 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:41.790 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:28:41.790 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:41.790 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:41.790 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:41.790 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:28:41.790 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:28:41.790 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:28:41.790 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:28:41.790 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:41.790 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:42.048 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:42.048 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:28:42.048 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:28:42.048 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:42.048 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:42.048 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:42.048 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:28:42.048 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:28:42.048 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:42.049 [2024-11-20 13:48:48.900272] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:28:42.049 [2024-11-20 13:48:48.902336] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:28:42.049 [2024-11-20 13:48:48.902401] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:28:42.049 [2024-11-20 13:48:48.902454] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:28:42.049 [2024-11-20 13:48:48.902513] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:28:42.049 [2024-11-20 13:48:48.902532] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:28:42.049 [2024-11-20 13:48:48.902549] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:28:42.049 [2024-11-20 13:48:48.902560] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:28:42.049 request: 00:28:42.049 { 00:28:42.049 "name": "raid_bdev1", 00:28:42.049 "raid_level": "raid1", 00:28:42.049 "base_bdevs": [ 00:28:42.049 "malloc1", 00:28:42.049 "malloc2", 00:28:42.049 "malloc3" 00:28:42.049 ], 00:28:42.049 "superblock": false, 00:28:42.049 "method": "bdev_raid_create", 00:28:42.049 "req_id": 1 00:28:42.049 } 00:28:42.049 Got JSON-RPC error response 00:28:42.049 response: 00:28:42.049 { 00:28:42.049 "code": -17, 00:28:42.049 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:28:42.049 } 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:42.049 [2024-11-20 13:48:48.940225] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:28:42.049 [2024-11-20 13:48:48.940292] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:42.049 [2024-11-20 13:48:48.940313] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:28:42.049 [2024-11-20 13:48:48.940323] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:42.049 [2024-11-20 13:48:48.942760] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:42.049 pt1 00:28:42.049 [2024-11-20 13:48:48.942924] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:28:42.049 [2024-11-20 13:48:48.943053] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:28:42.049 [2024-11-20 13:48:48.943111] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:42.049 "name": "raid_bdev1", 00:28:42.049 "uuid": "da5832c9-15ce-413d-9988-86373121ff09", 00:28:42.049 "strip_size_kb": 0, 00:28:42.049 "state": "configuring", 00:28:42.049 "raid_level": "raid1", 00:28:42.049 "superblock": true, 00:28:42.049 "num_base_bdevs": 3, 00:28:42.049 "num_base_bdevs_discovered": 1, 00:28:42.049 "num_base_bdevs_operational": 3, 00:28:42.049 "base_bdevs_list": [ 00:28:42.049 { 00:28:42.049 "name": "pt1", 00:28:42.049 "uuid": "00000000-0000-0000-0000-000000000001", 00:28:42.049 "is_configured": true, 00:28:42.049 "data_offset": 2048, 00:28:42.049 "data_size": 63488 00:28:42.049 }, 00:28:42.049 { 00:28:42.049 "name": null, 00:28:42.049 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:42.049 "is_configured": false, 00:28:42.049 "data_offset": 2048, 00:28:42.049 "data_size": 63488 00:28:42.049 }, 00:28:42.049 { 00:28:42.049 "name": null, 00:28:42.049 "uuid": "00000000-0000-0000-0000-000000000003", 00:28:42.049 "is_configured": false, 00:28:42.049 "data_offset": 2048, 00:28:42.049 "data_size": 63488 00:28:42.049 } 00:28:42.049 ] 00:28:42.049 }' 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:42.049 13:48:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:42.308 [2024-11-20 13:48:49.236313] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:28:42.308 [2024-11-20 13:48:49.236395] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:42.308 [2024-11-20 13:48:49.236420] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:28:42.308 [2024-11-20 13:48:49.236430] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:42.308 [2024-11-20 13:48:49.236919] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:42.308 [2024-11-20 13:48:49.236935] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:28:42.308 [2024-11-20 13:48:49.237061] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:28:42.308 [2024-11-20 13:48:49.237087] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:28:42.308 pt2 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:42.308 [2024-11-20 13:48:49.244310] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:42.308 "name": "raid_bdev1", 00:28:42.308 "uuid": "da5832c9-15ce-413d-9988-86373121ff09", 00:28:42.308 "strip_size_kb": 0, 00:28:42.308 "state": "configuring", 00:28:42.308 "raid_level": "raid1", 00:28:42.308 "superblock": true, 00:28:42.308 "num_base_bdevs": 3, 00:28:42.308 "num_base_bdevs_discovered": 1, 00:28:42.308 "num_base_bdevs_operational": 3, 00:28:42.308 "base_bdevs_list": [ 00:28:42.308 { 00:28:42.308 "name": "pt1", 00:28:42.308 "uuid": "00000000-0000-0000-0000-000000000001", 00:28:42.308 "is_configured": true, 00:28:42.308 "data_offset": 2048, 00:28:42.308 "data_size": 63488 00:28:42.308 }, 00:28:42.308 { 00:28:42.308 "name": null, 00:28:42.308 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:42.308 "is_configured": false, 00:28:42.308 "data_offset": 0, 00:28:42.308 "data_size": 63488 00:28:42.308 }, 00:28:42.308 { 00:28:42.308 "name": null, 00:28:42.308 "uuid": "00000000-0000-0000-0000-000000000003", 00:28:42.308 "is_configured": false, 00:28:42.308 "data_offset": 2048, 00:28:42.308 "data_size": 63488 00:28:42.308 } 00:28:42.308 ] 00:28:42.308 }' 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:42.308 13:48:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:42.567 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:28:42.567 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:28:42.567 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:28:42.567 13:48:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:42.567 13:48:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:42.567 [2024-11-20 13:48:49.568370] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:28:42.567 [2024-11-20 13:48:49.568462] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:42.568 [2024-11-20 13:48:49.568481] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:28:42.568 [2024-11-20 13:48:49.568493] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:42.568 [2024-11-20 13:48:49.569008] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:42.568 [2024-11-20 13:48:49.569030] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:28:42.568 [2024-11-20 13:48:49.569119] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:28:42.568 [2024-11-20 13:48:49.569153] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:28:42.568 pt2 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:42.568 [2024-11-20 13:48:49.576369] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:28:42.568 [2024-11-20 13:48:49.576433] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:42.568 [2024-11-20 13:48:49.576451] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:28:42.568 [2024-11-20 13:48:49.576461] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:42.568 [2024-11-20 13:48:49.576936] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:42.568 [2024-11-20 13:48:49.576965] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:28:42.568 [2024-11-20 13:48:49.577057] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:28:42.568 [2024-11-20 13:48:49.577084] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:28:42.568 [2024-11-20 13:48:49.577219] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:28:42.568 [2024-11-20 13:48:49.577234] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:28:42.568 [2024-11-20 13:48:49.577489] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:28:42.568 [2024-11-20 13:48:49.577645] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:28:42.568 [2024-11-20 13:48:49.577653] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:28:42.568 [2024-11-20 13:48:49.577790] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:42.568 pt3 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:42.568 "name": "raid_bdev1", 00:28:42.568 "uuid": "da5832c9-15ce-413d-9988-86373121ff09", 00:28:42.568 "strip_size_kb": 0, 00:28:42.568 "state": "online", 00:28:42.568 "raid_level": "raid1", 00:28:42.568 "superblock": true, 00:28:42.568 "num_base_bdevs": 3, 00:28:42.568 "num_base_bdevs_discovered": 3, 00:28:42.568 "num_base_bdevs_operational": 3, 00:28:42.568 "base_bdevs_list": [ 00:28:42.568 { 00:28:42.568 "name": "pt1", 00:28:42.568 "uuid": "00000000-0000-0000-0000-000000000001", 00:28:42.568 "is_configured": true, 00:28:42.568 "data_offset": 2048, 00:28:42.568 "data_size": 63488 00:28:42.568 }, 00:28:42.568 { 00:28:42.568 "name": "pt2", 00:28:42.568 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:42.568 "is_configured": true, 00:28:42.568 "data_offset": 2048, 00:28:42.568 "data_size": 63488 00:28:42.568 }, 00:28:42.568 { 00:28:42.568 "name": "pt3", 00:28:42.568 "uuid": "00000000-0000-0000-0000-000000000003", 00:28:42.568 "is_configured": true, 00:28:42.568 "data_offset": 2048, 00:28:42.568 "data_size": 63488 00:28:42.568 } 00:28:42.568 ] 00:28:42.568 }' 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:42.568 13:48:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:43.134 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:28:43.134 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:28:43.134 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:28:43.134 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:28:43.134 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:28:43.134 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:28:43.134 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:28:43.134 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:28:43.134 13:48:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:43.134 13:48:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:43.134 [2024-11-20 13:48:49.920794] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:43.134 13:48:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:43.134 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:43.134 "name": "raid_bdev1", 00:28:43.134 "aliases": [ 00:28:43.134 "da5832c9-15ce-413d-9988-86373121ff09" 00:28:43.134 ], 00:28:43.134 "product_name": "Raid Volume", 00:28:43.134 "block_size": 512, 00:28:43.134 "num_blocks": 63488, 00:28:43.134 "uuid": "da5832c9-15ce-413d-9988-86373121ff09", 00:28:43.134 "assigned_rate_limits": { 00:28:43.134 "rw_ios_per_sec": 0, 00:28:43.134 "rw_mbytes_per_sec": 0, 00:28:43.134 "r_mbytes_per_sec": 0, 00:28:43.134 "w_mbytes_per_sec": 0 00:28:43.134 }, 00:28:43.134 "claimed": false, 00:28:43.134 "zoned": false, 00:28:43.134 "supported_io_types": { 00:28:43.134 "read": true, 00:28:43.134 "write": true, 00:28:43.134 "unmap": false, 00:28:43.134 "flush": false, 00:28:43.134 "reset": true, 00:28:43.134 "nvme_admin": false, 00:28:43.134 "nvme_io": false, 00:28:43.134 "nvme_io_md": false, 00:28:43.134 "write_zeroes": true, 00:28:43.134 "zcopy": false, 00:28:43.134 "get_zone_info": false, 00:28:43.134 "zone_management": false, 00:28:43.134 "zone_append": false, 00:28:43.134 "compare": false, 00:28:43.134 "compare_and_write": false, 00:28:43.134 "abort": false, 00:28:43.134 "seek_hole": false, 00:28:43.134 "seek_data": false, 00:28:43.134 "copy": false, 00:28:43.134 "nvme_iov_md": false 00:28:43.134 }, 00:28:43.134 "memory_domains": [ 00:28:43.134 { 00:28:43.134 "dma_device_id": "system", 00:28:43.134 "dma_device_type": 1 00:28:43.134 }, 00:28:43.134 { 00:28:43.134 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:43.134 "dma_device_type": 2 00:28:43.134 }, 00:28:43.134 { 00:28:43.134 "dma_device_id": "system", 00:28:43.134 "dma_device_type": 1 00:28:43.134 }, 00:28:43.134 { 00:28:43.134 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:43.134 "dma_device_type": 2 00:28:43.134 }, 00:28:43.134 { 00:28:43.134 "dma_device_id": "system", 00:28:43.134 "dma_device_type": 1 00:28:43.134 }, 00:28:43.134 { 00:28:43.134 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:43.134 "dma_device_type": 2 00:28:43.134 } 00:28:43.134 ], 00:28:43.134 "driver_specific": { 00:28:43.134 "raid": { 00:28:43.134 "uuid": "da5832c9-15ce-413d-9988-86373121ff09", 00:28:43.134 "strip_size_kb": 0, 00:28:43.134 "state": "online", 00:28:43.134 "raid_level": "raid1", 00:28:43.134 "superblock": true, 00:28:43.134 "num_base_bdevs": 3, 00:28:43.134 "num_base_bdevs_discovered": 3, 00:28:43.134 "num_base_bdevs_operational": 3, 00:28:43.134 "base_bdevs_list": [ 00:28:43.134 { 00:28:43.134 "name": "pt1", 00:28:43.134 "uuid": "00000000-0000-0000-0000-000000000001", 00:28:43.134 "is_configured": true, 00:28:43.134 "data_offset": 2048, 00:28:43.134 "data_size": 63488 00:28:43.134 }, 00:28:43.134 { 00:28:43.134 "name": "pt2", 00:28:43.134 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:43.134 "is_configured": true, 00:28:43.134 "data_offset": 2048, 00:28:43.135 "data_size": 63488 00:28:43.135 }, 00:28:43.135 { 00:28:43.135 "name": "pt3", 00:28:43.135 "uuid": "00000000-0000-0000-0000-000000000003", 00:28:43.135 "is_configured": true, 00:28:43.135 "data_offset": 2048, 00:28:43.135 "data_size": 63488 00:28:43.135 } 00:28:43.135 ] 00:28:43.135 } 00:28:43.135 } 00:28:43.135 }' 00:28:43.135 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:28:43.135 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:28:43.135 pt2 00:28:43.135 pt3' 00:28:43.135 13:48:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:28:43.135 [2024-11-20 13:48:50.096804] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' da5832c9-15ce-413d-9988-86373121ff09 '!=' da5832c9-15ce-413d-9988-86373121ff09 ']' 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:43.135 [2024-11-20 13:48:50.132563] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:43.135 "name": "raid_bdev1", 00:28:43.135 "uuid": "da5832c9-15ce-413d-9988-86373121ff09", 00:28:43.135 "strip_size_kb": 0, 00:28:43.135 "state": "online", 00:28:43.135 "raid_level": "raid1", 00:28:43.135 "superblock": true, 00:28:43.135 "num_base_bdevs": 3, 00:28:43.135 "num_base_bdevs_discovered": 2, 00:28:43.135 "num_base_bdevs_operational": 2, 00:28:43.135 "base_bdevs_list": [ 00:28:43.135 { 00:28:43.135 "name": null, 00:28:43.135 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:43.135 "is_configured": false, 00:28:43.135 "data_offset": 0, 00:28:43.135 "data_size": 63488 00:28:43.135 }, 00:28:43.135 { 00:28:43.135 "name": "pt2", 00:28:43.135 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:43.135 "is_configured": true, 00:28:43.135 "data_offset": 2048, 00:28:43.135 "data_size": 63488 00:28:43.135 }, 00:28:43.135 { 00:28:43.135 "name": "pt3", 00:28:43.135 "uuid": "00000000-0000-0000-0000-000000000003", 00:28:43.135 "is_configured": true, 00:28:43.135 "data_offset": 2048, 00:28:43.135 "data_size": 63488 00:28:43.135 } 00:28:43.135 ] 00:28:43.135 }' 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:43.135 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:43.395 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:28:43.395 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:43.395 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:43.395 [2024-11-20 13:48:50.436584] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:28:43.395 [2024-11-20 13:48:50.436626] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:28:43.395 [2024-11-20 13:48:50.436705] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:43.395 [2024-11-20 13:48:50.436773] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:43.395 [2024-11-20 13:48:50.436788] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:28:43.395 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:43.395 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:28:43.395 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:43.395 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:43.395 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:43.653 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:43.653 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:28:43.653 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:28:43.653 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:28:43.653 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:28:43.653 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:28:43.653 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:43.653 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:43.653 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:43.653 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:28:43.653 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:28:43.653 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:28:43.653 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:43.653 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:43.653 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:43.653 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:28:43.653 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:28:43.653 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:28:43.653 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:28:43.653 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:28:43.653 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:43.653 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:43.653 [2024-11-20 13:48:50.508572] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:28:43.653 [2024-11-20 13:48:50.508650] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:43.653 [2024-11-20 13:48:50.508669] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:28:43.653 [2024-11-20 13:48:50.508680] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:43.654 [2024-11-20 13:48:50.511186] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:43.654 [2024-11-20 13:48:50.511232] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:28:43.654 [2024-11-20 13:48:50.511323] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:28:43.654 [2024-11-20 13:48:50.511376] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:28:43.654 pt2 00:28:43.654 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:43.654 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:28:43.654 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:43.654 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:43.654 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:43.654 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:43.654 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:28:43.654 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:43.654 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:43.654 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:43.654 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:43.654 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:43.654 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:43.654 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:43.654 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:43.654 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:43.654 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:43.654 "name": "raid_bdev1", 00:28:43.654 "uuid": "da5832c9-15ce-413d-9988-86373121ff09", 00:28:43.654 "strip_size_kb": 0, 00:28:43.654 "state": "configuring", 00:28:43.654 "raid_level": "raid1", 00:28:43.654 "superblock": true, 00:28:43.654 "num_base_bdevs": 3, 00:28:43.654 "num_base_bdevs_discovered": 1, 00:28:43.654 "num_base_bdevs_operational": 2, 00:28:43.654 "base_bdevs_list": [ 00:28:43.654 { 00:28:43.654 "name": null, 00:28:43.654 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:43.654 "is_configured": false, 00:28:43.654 "data_offset": 2048, 00:28:43.654 "data_size": 63488 00:28:43.654 }, 00:28:43.654 { 00:28:43.654 "name": "pt2", 00:28:43.654 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:43.654 "is_configured": true, 00:28:43.654 "data_offset": 2048, 00:28:43.654 "data_size": 63488 00:28:43.654 }, 00:28:43.654 { 00:28:43.654 "name": null, 00:28:43.654 "uuid": "00000000-0000-0000-0000-000000000003", 00:28:43.654 "is_configured": false, 00:28:43.654 "data_offset": 2048, 00:28:43.654 "data_size": 63488 00:28:43.654 } 00:28:43.654 ] 00:28:43.654 }' 00:28:43.654 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:43.654 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:43.913 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:28:43.913 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:28:43.913 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=2 00:28:43.913 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:28:43.913 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:43.913 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:43.913 [2024-11-20 13:48:50.828676] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:28:43.913 [2024-11-20 13:48:50.828764] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:43.913 [2024-11-20 13:48:50.828785] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:28:43.913 [2024-11-20 13:48:50.828796] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:43.913 [2024-11-20 13:48:50.829324] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:43.913 [2024-11-20 13:48:50.829369] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:28:43.913 [2024-11-20 13:48:50.829459] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:28:43.913 [2024-11-20 13:48:50.829488] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:28:43.913 [2024-11-20 13:48:50.829607] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:28:43.913 [2024-11-20 13:48:50.829633] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:28:43.913 [2024-11-20 13:48:50.829897] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:28:43.913 [2024-11-20 13:48:50.830066] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:28:43.913 [2024-11-20 13:48:50.830080] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:28:43.913 [2024-11-20 13:48:50.830223] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:43.913 pt3 00:28:43.913 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:43.913 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:28:43.913 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:43.913 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:43.913 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:43.913 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:43.913 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:28:43.913 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:43.913 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:43.913 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:43.913 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:43.913 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:43.913 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:43.913 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:43.913 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:43.913 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:43.913 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:43.913 "name": "raid_bdev1", 00:28:43.913 "uuid": "da5832c9-15ce-413d-9988-86373121ff09", 00:28:43.913 "strip_size_kb": 0, 00:28:43.913 "state": "online", 00:28:43.913 "raid_level": "raid1", 00:28:43.913 "superblock": true, 00:28:43.913 "num_base_bdevs": 3, 00:28:43.913 "num_base_bdevs_discovered": 2, 00:28:43.913 "num_base_bdevs_operational": 2, 00:28:43.913 "base_bdevs_list": [ 00:28:43.913 { 00:28:43.913 "name": null, 00:28:43.913 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:43.913 "is_configured": false, 00:28:43.913 "data_offset": 2048, 00:28:43.913 "data_size": 63488 00:28:43.913 }, 00:28:43.913 { 00:28:43.913 "name": "pt2", 00:28:43.913 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:43.913 "is_configured": true, 00:28:43.913 "data_offset": 2048, 00:28:43.913 "data_size": 63488 00:28:43.913 }, 00:28:43.913 { 00:28:43.913 "name": "pt3", 00:28:43.913 "uuid": "00000000-0000-0000-0000-000000000003", 00:28:43.913 "is_configured": true, 00:28:43.913 "data_offset": 2048, 00:28:43.913 "data_size": 63488 00:28:43.913 } 00:28:43.913 ] 00:28:43.913 }' 00:28:43.913 13:48:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:43.913 13:48:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:44.177 [2024-11-20 13:48:51.152722] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:28:44.177 [2024-11-20 13:48:51.152761] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:28:44.177 [2024-11-20 13:48:51.152841] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:44.177 [2024-11-20 13:48:51.152915] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:44.177 [2024-11-20 13:48:51.152925] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 3 -gt 2 ']' 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@534 -- # i=2 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt3 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:44.177 [2024-11-20 13:48:51.204776] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:28:44.177 [2024-11-20 13:48:51.204855] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:44.177 [2024-11-20 13:48:51.204874] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:28:44.177 [2024-11-20 13:48:51.204884] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:44.177 [2024-11-20 13:48:51.207345] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:44.177 [2024-11-20 13:48:51.207387] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:28:44.177 [2024-11-20 13:48:51.207482] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:28:44.177 [2024-11-20 13:48:51.207534] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:28:44.177 [2024-11-20 13:48:51.207663] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:28:44.177 [2024-11-20 13:48:51.207684] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:28:44.177 [2024-11-20 13:48:51.207703] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:28:44.177 [2024-11-20 13:48:51.207757] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:28:44.177 pt1 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 3 -gt 2 ']' 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:44.177 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.436 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:44.436 "name": "raid_bdev1", 00:28:44.436 "uuid": "da5832c9-15ce-413d-9988-86373121ff09", 00:28:44.436 "strip_size_kb": 0, 00:28:44.436 "state": "configuring", 00:28:44.436 "raid_level": "raid1", 00:28:44.436 "superblock": true, 00:28:44.436 "num_base_bdevs": 3, 00:28:44.436 "num_base_bdevs_discovered": 1, 00:28:44.436 "num_base_bdevs_operational": 2, 00:28:44.436 "base_bdevs_list": [ 00:28:44.436 { 00:28:44.436 "name": null, 00:28:44.436 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:44.436 "is_configured": false, 00:28:44.436 "data_offset": 2048, 00:28:44.436 "data_size": 63488 00:28:44.436 }, 00:28:44.436 { 00:28:44.436 "name": "pt2", 00:28:44.436 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:44.436 "is_configured": true, 00:28:44.436 "data_offset": 2048, 00:28:44.436 "data_size": 63488 00:28:44.436 }, 00:28:44.436 { 00:28:44.436 "name": null, 00:28:44.436 "uuid": "00000000-0000-0000-0000-000000000003", 00:28:44.436 "is_configured": false, 00:28:44.436 "data_offset": 2048, 00:28:44.436 "data_size": 63488 00:28:44.436 } 00:28:44.436 ] 00:28:44.436 }' 00:28:44.436 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:44.436 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:44.694 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:28:44.694 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.694 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:28:44.694 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:44.694 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.694 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:28:44.694 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:28:44.694 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.694 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:44.694 [2024-11-20 13:48:51.556862] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:28:44.694 [2024-11-20 13:48:51.556939] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:44.694 [2024-11-20 13:48:51.556961] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:28:44.694 [2024-11-20 13:48:51.556971] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:44.694 [2024-11-20 13:48:51.557468] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:44.694 [2024-11-20 13:48:51.557489] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:28:44.694 [2024-11-20 13:48:51.557570] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:28:44.694 [2024-11-20 13:48:51.557592] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:28:44.694 [2024-11-20 13:48:51.557710] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:28:44.694 [2024-11-20 13:48:51.557719] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:28:44.694 [2024-11-20 13:48:51.557974] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:28:44.694 [2024-11-20 13:48:51.558143] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:28:44.694 [2024-11-20 13:48:51.558156] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:28:44.694 [2024-11-20 13:48:51.558288] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:44.694 pt3 00:28:44.694 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.694 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:28:44.694 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:44.694 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:44.694 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:44.694 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:44.694 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:28:44.694 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:44.694 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:44.694 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:44.694 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:44.694 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:44.694 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.694 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:44.694 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:44.694 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.694 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:44.694 "name": "raid_bdev1", 00:28:44.694 "uuid": "da5832c9-15ce-413d-9988-86373121ff09", 00:28:44.694 "strip_size_kb": 0, 00:28:44.694 "state": "online", 00:28:44.694 "raid_level": "raid1", 00:28:44.694 "superblock": true, 00:28:44.694 "num_base_bdevs": 3, 00:28:44.694 "num_base_bdevs_discovered": 2, 00:28:44.694 "num_base_bdevs_operational": 2, 00:28:44.695 "base_bdevs_list": [ 00:28:44.695 { 00:28:44.695 "name": null, 00:28:44.695 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:44.695 "is_configured": false, 00:28:44.695 "data_offset": 2048, 00:28:44.695 "data_size": 63488 00:28:44.695 }, 00:28:44.695 { 00:28:44.695 "name": "pt2", 00:28:44.695 "uuid": "00000000-0000-0000-0000-000000000002", 00:28:44.695 "is_configured": true, 00:28:44.695 "data_offset": 2048, 00:28:44.695 "data_size": 63488 00:28:44.695 }, 00:28:44.695 { 00:28:44.695 "name": "pt3", 00:28:44.695 "uuid": "00000000-0000-0000-0000-000000000003", 00:28:44.695 "is_configured": true, 00:28:44.695 "data_offset": 2048, 00:28:44.695 "data_size": 63488 00:28:44.695 } 00:28:44.695 ] 00:28:44.695 }' 00:28:44.695 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:44.695 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:44.952 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:28:44.952 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.952 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:44.952 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:28:44.952 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.952 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:28:44.952 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:28:44.952 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:28:44.952 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:44.952 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:44.953 [2024-11-20 13:48:51.933265] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:44.953 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:44.953 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' da5832c9-15ce-413d-9988-86373121ff09 '!=' da5832c9-15ce-413d-9988-86373121ff09 ']' 00:28:44.953 13:48:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 66919 00:28:44.953 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 66919 ']' 00:28:44.953 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 66919 00:28:44.953 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:28:44.953 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:28:44.953 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 66919 00:28:44.953 killing process with pid 66919 00:28:44.953 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:28:44.953 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:28:44.953 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 66919' 00:28:44.953 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 66919 00:28:44.953 [2024-11-20 13:48:51.981890] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:28:44.953 13:48:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 66919 00:28:44.953 [2024-11-20 13:48:51.982019] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:44.953 [2024-11-20 13:48:51.982096] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:44.953 [2024-11-20 13:48:51.982109] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:28:45.211 [2024-11-20 13:48:52.171726] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:28:45.777 ************************************ 00:28:45.777 END TEST raid_superblock_test 00:28:45.777 ************************************ 00:28:45.777 13:48:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:28:45.777 00:28:45.777 real 0m5.643s 00:28:45.777 user 0m8.829s 00:28:45.777 sys 0m0.954s 00:28:45.777 13:48:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:28:45.777 13:48:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:28:46.034 13:48:52 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 3 read 00:28:46.034 13:48:52 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:28:46.034 13:48:52 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:28:46.034 13:48:52 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:28:46.034 ************************************ 00:28:46.034 START TEST raid_read_error_test 00:28:46.034 ************************************ 00:28:46.034 13:48:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 3 read 00:28:46.034 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:28:46.034 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:28:46.034 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:28:46.034 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:28:46.034 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:28:46.034 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:28:46.034 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:28:46.035 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.yF9ak35WGn 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=67342 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 67342 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 67342 ']' 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:46.035 13:48:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:28:46.035 [2024-11-20 13:48:52.937846] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:28:46.035 [2024-11-20 13:48:52.937971] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid67342 ] 00:28:46.292 [2024-11-20 13:48:53.095937] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:46.292 [2024-11-20 13:48:53.220156] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:28:46.551 [2024-11-20 13:48:53.370478] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:46.551 [2024-11-20 13:48:53.370536] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:46.808 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:28:46.808 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:28:46.808 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:28:46.808 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:28:46.808 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:46.808 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:46.808 BaseBdev1_malloc 00:28:46.808 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:46.808 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:28:46.808 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:46.808 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:46.808 true 00:28:46.808 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:46.808 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:28:46.808 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:46.808 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:46.808 [2024-11-20 13:48:53.839006] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:28:46.808 [2024-11-20 13:48:53.839076] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:46.808 [2024-11-20 13:48:53.839101] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:28:46.808 [2024-11-20 13:48:53.839112] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:46.808 [2024-11-20 13:48:53.841501] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:46.808 [2024-11-20 13:48:53.841720] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:28:46.808 BaseBdev1 00:28:46.808 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:46.808 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:28:46.808 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:28:46.808 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:46.808 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:47.066 BaseBdev2_malloc 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:47.066 true 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:47.066 [2024-11-20 13:48:53.889415] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:28:47.066 [2024-11-20 13:48:53.889480] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:47.066 [2024-11-20 13:48:53.889502] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:28:47.066 [2024-11-20 13:48:53.889515] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:47.066 [2024-11-20 13:48:53.891865] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:47.066 [2024-11-20 13:48:53.891902] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:28:47.066 BaseBdev2 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:47.066 BaseBdev3_malloc 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:47.066 true 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:47.066 [2024-11-20 13:48:53.948365] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:28:47.066 [2024-11-20 13:48:53.948436] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:47.066 [2024-11-20 13:48:53.948460] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:28:47.066 [2024-11-20 13:48:53.948471] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:47.066 [2024-11-20 13:48:53.950884] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:47.066 [2024-11-20 13:48:53.950926] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:28:47.066 BaseBdev3 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:47.066 [2024-11-20 13:48:53.956434] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:47.066 [2024-11-20 13:48:53.958549] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:28:47.066 [2024-11-20 13:48:53.958631] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:28:47.066 [2024-11-20 13:48:53.958852] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:28:47.066 [2024-11-20 13:48:53.958864] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:28:47.066 [2024-11-20 13:48:53.959181] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006560 00:28:47.066 [2024-11-20 13:48:53.959345] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:28:47.066 [2024-11-20 13:48:53.959357] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:28:47.066 [2024-11-20 13:48:53.959519] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:47.066 "name": "raid_bdev1", 00:28:47.066 "uuid": "e131879d-53ce-45c4-b36b-b461da68cee2", 00:28:47.066 "strip_size_kb": 0, 00:28:47.066 "state": "online", 00:28:47.066 "raid_level": "raid1", 00:28:47.066 "superblock": true, 00:28:47.066 "num_base_bdevs": 3, 00:28:47.066 "num_base_bdevs_discovered": 3, 00:28:47.066 "num_base_bdevs_operational": 3, 00:28:47.066 "base_bdevs_list": [ 00:28:47.066 { 00:28:47.066 "name": "BaseBdev1", 00:28:47.066 "uuid": "11d57a73-9d9c-5494-a7ff-b7a49a5ae28b", 00:28:47.066 "is_configured": true, 00:28:47.066 "data_offset": 2048, 00:28:47.066 "data_size": 63488 00:28:47.066 }, 00:28:47.066 { 00:28:47.066 "name": "BaseBdev2", 00:28:47.066 "uuid": "dc5f3296-2b6e-50c3-84af-6b37194c00dd", 00:28:47.066 "is_configured": true, 00:28:47.066 "data_offset": 2048, 00:28:47.066 "data_size": 63488 00:28:47.066 }, 00:28:47.066 { 00:28:47.066 "name": "BaseBdev3", 00:28:47.066 "uuid": "e8af3505-bbcc-5ec2-baa7-1595f131ed28", 00:28:47.066 "is_configured": true, 00:28:47.066 "data_offset": 2048, 00:28:47.066 "data_size": 63488 00:28:47.066 } 00:28:47.066 ] 00:28:47.066 }' 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:47.066 13:48:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:47.324 13:48:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:28:47.324 13:48:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:28:47.582 [2024-11-20 13:48:54.397633] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006700 00:28:48.516 13:48:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:28:48.516 13:48:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:48.516 13:48:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:48.516 13:48:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:48.516 13:48:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:28:48.516 13:48:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:28:48.516 13:48:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:28:48.516 13:48:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:28:48.516 13:48:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:28:48.516 13:48:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:48.516 13:48:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:48.516 13:48:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:48.516 13:48:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:48.516 13:48:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:48.516 13:48:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:48.516 13:48:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:48.516 13:48:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:48.516 13:48:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:48.516 13:48:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:48.516 13:48:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:48.516 13:48:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:48.516 13:48:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:48.516 13:48:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:48.516 13:48:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:48.516 "name": "raid_bdev1", 00:28:48.516 "uuid": "e131879d-53ce-45c4-b36b-b461da68cee2", 00:28:48.516 "strip_size_kb": 0, 00:28:48.516 "state": "online", 00:28:48.516 "raid_level": "raid1", 00:28:48.516 "superblock": true, 00:28:48.516 "num_base_bdevs": 3, 00:28:48.516 "num_base_bdevs_discovered": 3, 00:28:48.516 "num_base_bdevs_operational": 3, 00:28:48.516 "base_bdevs_list": [ 00:28:48.516 { 00:28:48.516 "name": "BaseBdev1", 00:28:48.516 "uuid": "11d57a73-9d9c-5494-a7ff-b7a49a5ae28b", 00:28:48.516 "is_configured": true, 00:28:48.516 "data_offset": 2048, 00:28:48.516 "data_size": 63488 00:28:48.516 }, 00:28:48.516 { 00:28:48.516 "name": "BaseBdev2", 00:28:48.516 "uuid": "dc5f3296-2b6e-50c3-84af-6b37194c00dd", 00:28:48.516 "is_configured": true, 00:28:48.516 "data_offset": 2048, 00:28:48.516 "data_size": 63488 00:28:48.516 }, 00:28:48.516 { 00:28:48.516 "name": "BaseBdev3", 00:28:48.516 "uuid": "e8af3505-bbcc-5ec2-baa7-1595f131ed28", 00:28:48.516 "is_configured": true, 00:28:48.516 "data_offset": 2048, 00:28:48.516 "data_size": 63488 00:28:48.516 } 00:28:48.516 ] 00:28:48.516 }' 00:28:48.516 13:48:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:48.516 13:48:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:48.773 13:48:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:28:48.773 13:48:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:48.773 13:48:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:48.773 [2024-11-20 13:48:55.631210] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:28:48.773 [2024-11-20 13:48:55.631419] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:28:48.773 [2024-11-20 13:48:55.634641] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:48.773 [2024-11-20 13:48:55.634795] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:48.773 [2024-11-20 13:48:55.634920] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:48.773 [2024-11-20 13:48:55.634931] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:28:48.773 { 00:28:48.773 "results": [ 00:28:48.773 { 00:28:48.773 "job": "raid_bdev1", 00:28:48.773 "core_mask": "0x1", 00:28:48.773 "workload": "randrw", 00:28:48.773 "percentage": 50, 00:28:48.773 "status": "finished", 00:28:48.773 "queue_depth": 1, 00:28:48.773 "io_size": 131072, 00:28:48.773 "runtime": 1.231812, 00:28:48.773 "iops": 12031.056687221751, 00:28:48.773 "mibps": 1503.882085902719, 00:28:48.773 "io_failed": 0, 00:28:48.773 "io_timeout": 0, 00:28:48.773 "avg_latency_us": 79.8340371639157, 00:28:48.773 "min_latency_us": 30.12923076923077, 00:28:48.773 "max_latency_us": 1764.4307692307693 00:28:48.773 } 00:28:48.773 ], 00:28:48.774 "core_count": 1 00:28:48.774 } 00:28:48.774 13:48:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:48.774 13:48:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 67342 00:28:48.774 13:48:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 67342 ']' 00:28:48.774 13:48:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 67342 00:28:48.774 13:48:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:28:48.774 13:48:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:28:48.774 13:48:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 67342 00:28:48.774 killing process with pid 67342 00:28:48.774 13:48:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:28:48.774 13:48:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:28:48.774 13:48:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 67342' 00:28:48.774 13:48:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 67342 00:28:48.774 13:48:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 67342 00:28:48.774 [2024-11-20 13:48:55.664418] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:28:48.774 [2024-11-20 13:48:55.825036] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:28:49.706 13:48:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:28:49.706 13:48:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:28:49.706 13:48:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.yF9ak35WGn 00:28:49.706 ************************************ 00:28:49.706 END TEST raid_read_error_test 00:28:49.706 ************************************ 00:28:49.706 13:48:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:28:49.706 13:48:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:28:49.706 13:48:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:28:49.706 13:48:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:28:49.706 13:48:56 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:28:49.706 00:28:49.706 real 0m3.797s 00:28:49.706 user 0m4.466s 00:28:49.706 sys 0m0.448s 00:28:49.706 13:48:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:28:49.706 13:48:56 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:49.706 13:48:56 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 3 write 00:28:49.706 13:48:56 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:28:49.706 13:48:56 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:28:49.706 13:48:56 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:28:49.706 ************************************ 00:28:49.706 START TEST raid_write_error_test 00:28:49.706 ************************************ 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 3 write 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:28:49.706 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.lYrnlcmNQM 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=67477 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 67477 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 67477 ']' 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:49.706 13:48:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:28:49.965 [2024-11-20 13:48:56.791414] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:28:49.965 [2024-11-20 13:48:56.791563] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid67477 ] 00:28:49.965 [2024-11-20 13:48:56.952000] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:50.222 [2024-11-20 13:48:57.073685] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:28:50.222 [2024-11-20 13:48:57.224739] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:50.222 [2024-11-20 13:48:57.224796] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:50.787 BaseBdev1_malloc 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:50.787 true 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:50.787 [2024-11-20 13:48:57.694521] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:28:50.787 [2024-11-20 13:48:57.694584] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:50.787 [2024-11-20 13:48:57.694605] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:28:50.787 [2024-11-20 13:48:57.694616] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:50.787 [2024-11-20 13:48:57.696975] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:50.787 [2024-11-20 13:48:57.697024] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:28:50.787 BaseBdev1 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:50.787 BaseBdev2_malloc 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:50.787 true 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:50.787 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:50.787 [2024-11-20 13:48:57.740902] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:28:50.787 [2024-11-20 13:48:57.740958] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:50.787 [2024-11-20 13:48:57.740975] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:28:50.787 [2024-11-20 13:48:57.741002] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:50.788 [2024-11-20 13:48:57.743276] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:50.788 [2024-11-20 13:48:57.743309] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:28:50.788 BaseBdev2 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:50.788 BaseBdev3_malloc 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:50.788 true 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:50.788 [2024-11-20 13:48:57.803835] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:28:50.788 [2024-11-20 13:48:57.804018] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:50.788 [2024-11-20 13:48:57.804045] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:28:50.788 [2024-11-20 13:48:57.804058] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:50.788 [2024-11-20 13:48:57.806406] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:50.788 [2024-11-20 13:48:57.806443] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:28:50.788 BaseBdev3 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:50.788 [2024-11-20 13:48:57.811898] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:50.788 [2024-11-20 13:48:57.813957] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:28:50.788 [2024-11-20 13:48:57.814054] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:28:50.788 [2024-11-20 13:48:57.814269] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:28:50.788 [2024-11-20 13:48:57.814290] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:28:50.788 [2024-11-20 13:48:57.814560] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006560 00:28:50.788 [2024-11-20 13:48:57.814732] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:28:50.788 [2024-11-20 13:48:57.814744] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:28:50.788 [2024-11-20 13:48:57.814890] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:50.788 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:51.046 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:51.046 "name": "raid_bdev1", 00:28:51.046 "uuid": "b5dfa780-df74-473d-b7d5-3e54a75d1e79", 00:28:51.046 "strip_size_kb": 0, 00:28:51.046 "state": "online", 00:28:51.046 "raid_level": "raid1", 00:28:51.046 "superblock": true, 00:28:51.046 "num_base_bdevs": 3, 00:28:51.046 "num_base_bdevs_discovered": 3, 00:28:51.046 "num_base_bdevs_operational": 3, 00:28:51.046 "base_bdevs_list": [ 00:28:51.046 { 00:28:51.046 "name": "BaseBdev1", 00:28:51.046 "uuid": "8f64f1bd-715c-5572-8648-f9b4551c1044", 00:28:51.046 "is_configured": true, 00:28:51.046 "data_offset": 2048, 00:28:51.046 "data_size": 63488 00:28:51.046 }, 00:28:51.046 { 00:28:51.046 "name": "BaseBdev2", 00:28:51.046 "uuid": "77a9e5c2-f329-5d79-b36b-5e3374beb95b", 00:28:51.046 "is_configured": true, 00:28:51.046 "data_offset": 2048, 00:28:51.046 "data_size": 63488 00:28:51.046 }, 00:28:51.046 { 00:28:51.046 "name": "BaseBdev3", 00:28:51.046 "uuid": "afa38cb5-0fcf-5cb4-bef0-6ecf22a5bceb", 00:28:51.046 "is_configured": true, 00:28:51.046 "data_offset": 2048, 00:28:51.046 "data_size": 63488 00:28:51.046 } 00:28:51.046 ] 00:28:51.046 }' 00:28:51.046 13:48:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:51.046 13:48:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:51.304 13:48:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:28:51.304 13:48:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:28:51.304 [2024-11-20 13:48:58.217076] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006700 00:28:52.239 13:48:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:28:52.240 13:48:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:52.240 13:48:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:52.240 [2024-11-20 13:48:59.138450] bdev_raid.c:2276:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:28:52.240 [2024-11-20 13:48:59.138507] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:28:52.240 [2024-11-20 13:48:59.138730] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006700 00:28:52.240 13:48:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:52.240 13:48:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:28:52.240 13:48:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:28:52.240 13:48:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:28:52.240 13:48:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=2 00:28:52.240 13:48:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:28:52.240 13:48:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:52.240 13:48:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:52.240 13:48:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:52.240 13:48:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:52.240 13:48:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:28:52.240 13:48:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:52.240 13:48:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:52.240 13:48:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:52.240 13:48:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:52.240 13:48:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:52.240 13:48:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:52.240 13:48:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:52.240 13:48:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:52.241 13:48:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:52.241 13:48:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:52.241 "name": "raid_bdev1", 00:28:52.241 "uuid": "b5dfa780-df74-473d-b7d5-3e54a75d1e79", 00:28:52.241 "strip_size_kb": 0, 00:28:52.241 "state": "online", 00:28:52.241 "raid_level": "raid1", 00:28:52.241 "superblock": true, 00:28:52.241 "num_base_bdevs": 3, 00:28:52.241 "num_base_bdevs_discovered": 2, 00:28:52.241 "num_base_bdevs_operational": 2, 00:28:52.241 "base_bdevs_list": [ 00:28:52.241 { 00:28:52.241 "name": null, 00:28:52.241 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:52.241 "is_configured": false, 00:28:52.241 "data_offset": 0, 00:28:52.241 "data_size": 63488 00:28:52.241 }, 00:28:52.241 { 00:28:52.241 "name": "BaseBdev2", 00:28:52.241 "uuid": "77a9e5c2-f329-5d79-b36b-5e3374beb95b", 00:28:52.241 "is_configured": true, 00:28:52.241 "data_offset": 2048, 00:28:52.241 "data_size": 63488 00:28:52.241 }, 00:28:52.241 { 00:28:52.241 "name": "BaseBdev3", 00:28:52.241 "uuid": "afa38cb5-0fcf-5cb4-bef0-6ecf22a5bceb", 00:28:52.241 "is_configured": true, 00:28:52.241 "data_offset": 2048, 00:28:52.241 "data_size": 63488 00:28:52.241 } 00:28:52.241 ] 00:28:52.241 }' 00:28:52.241 13:48:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:52.241 13:48:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:52.504 13:48:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:28:52.504 13:48:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:52.504 13:48:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:52.504 [2024-11-20 13:48:59.484607] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:28:52.504 [2024-11-20 13:48:59.484647] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:28:52.504 [2024-11-20 13:48:59.487929] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:52.504 [2024-11-20 13:48:59.488073] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:52.504 [2024-11-20 13:48:59.488192] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:52.504 [2024-11-20 13:48:59.488281] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:28:52.504 13:48:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:52.504 { 00:28:52.504 "results": [ 00:28:52.504 { 00:28:52.504 "job": "raid_bdev1", 00:28:52.504 "core_mask": "0x1", 00:28:52.504 "workload": "randrw", 00:28:52.504 "percentage": 50, 00:28:52.504 "status": "finished", 00:28:52.504 "queue_depth": 1, 00:28:52.504 "io_size": 131072, 00:28:52.504 "runtime": 1.265495, 00:28:52.504 "iops": 13558.33092979427, 00:28:52.504 "mibps": 1694.7913662242838, 00:28:52.504 "io_failed": 0, 00:28:52.504 "io_timeout": 0, 00:28:52.504 "avg_latency_us": 70.4579705362827, 00:28:52.504 "min_latency_us": 30.12923076923077, 00:28:52.504 "max_latency_us": 1726.6215384615384 00:28:52.504 } 00:28:52.504 ], 00:28:52.504 "core_count": 1 00:28:52.504 } 00:28:52.505 13:48:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 67477 00:28:52.505 13:48:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 67477 ']' 00:28:52.505 13:48:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 67477 00:28:52.505 13:48:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:28:52.505 13:48:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:28:52.505 13:48:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 67477 00:28:52.505 13:48:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:28:52.505 killing process with pid 67477 00:28:52.505 13:48:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:28:52.505 13:48:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 67477' 00:28:52.505 13:48:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 67477 00:28:52.505 [2024-11-20 13:48:59.519281] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:28:52.505 13:48:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 67477 00:28:52.763 [2024-11-20 13:48:59.651124] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:28:53.329 13:49:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.lYrnlcmNQM 00:28:53.329 13:49:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:28:53.329 13:49:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:28:53.329 13:49:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:28:53.329 13:49:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:28:53.329 13:49:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:28:53.329 13:49:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:28:53.329 13:49:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:28:53.329 00:28:53.329 real 0m3.611s 00:28:53.329 user 0m4.281s 00:28:53.329 sys 0m0.439s 00:28:53.329 13:49:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:28:53.329 ************************************ 00:28:53.329 END TEST raid_write_error_test 00:28:53.329 ************************************ 00:28:53.329 13:49:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:28:53.329 13:49:00 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:28:53.329 13:49:00 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:28:53.329 13:49:00 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 4 false 00:28:53.329 13:49:00 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:28:53.329 13:49:00 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:28:53.329 13:49:00 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:28:53.587 ************************************ 00:28:53.587 START TEST raid_state_function_test 00:28:53.587 ************************************ 00:28:53.587 13:49:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 4 false 00:28:53.587 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:28:53.587 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:28:53.587 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:28:53.587 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:28:53.587 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:28:53.587 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:28:53.587 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:28:53.587 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:28:53.587 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:28:53.587 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:28:53.587 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:28:53.587 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:28:53.587 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:28:53.587 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:28:53.587 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:28:53.587 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:28:53.587 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:28:53.587 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:28:53.587 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:28:53.587 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:28:53.587 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:28:53.588 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:28:53.588 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:28:53.588 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:28:53.588 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:28:53.588 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:28:53.588 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:28:53.588 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:28:53.588 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:28:53.588 Process raid pid: 67609 00:28:53.588 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:28:53.588 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=67609 00:28:53.588 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 67609' 00:28:53.588 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 67609 00:28:53.588 13:49:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 67609 ']' 00:28:53.588 13:49:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:28:53.588 13:49:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:28:53.588 13:49:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:28:53.588 13:49:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:28:53.588 13:49:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:28:53.588 13:49:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:53.588 [2024-11-20 13:49:00.466560] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:28:53.588 [2024-11-20 13:49:00.466685] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:28:53.588 [2024-11-20 13:49:00.622364] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:53.845 [2024-11-20 13:49:00.727664] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:28:53.845 [2024-11-20 13:49:00.853674] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:53.845 [2024-11-20 13:49:00.853725] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:54.411 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:28:54.411 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:28:54.411 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:28:54.411 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:54.411 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:54.411 [2024-11-20 13:49:01.287724] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:28:54.411 [2024-11-20 13:49:01.287795] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:28:54.411 [2024-11-20 13:49:01.287805] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:28:54.411 [2024-11-20 13:49:01.287813] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:28:54.411 [2024-11-20 13:49:01.287818] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:28:54.411 [2024-11-20 13:49:01.287833] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:28:54.411 [2024-11-20 13:49:01.287838] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:28:54.411 [2024-11-20 13:49:01.287845] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:28:54.411 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:54.411 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:28:54.411 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:54.412 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:54.412 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:28:54.412 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:54.412 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:28:54.412 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:54.412 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:54.412 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:54.412 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:54.412 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:54.412 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:54.412 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:54.412 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:54.412 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:54.412 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:54.412 "name": "Existed_Raid", 00:28:54.412 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:54.412 "strip_size_kb": 64, 00:28:54.412 "state": "configuring", 00:28:54.412 "raid_level": "raid0", 00:28:54.412 "superblock": false, 00:28:54.412 "num_base_bdevs": 4, 00:28:54.412 "num_base_bdevs_discovered": 0, 00:28:54.412 "num_base_bdevs_operational": 4, 00:28:54.412 "base_bdevs_list": [ 00:28:54.412 { 00:28:54.412 "name": "BaseBdev1", 00:28:54.412 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:54.412 "is_configured": false, 00:28:54.412 "data_offset": 0, 00:28:54.412 "data_size": 0 00:28:54.412 }, 00:28:54.412 { 00:28:54.412 "name": "BaseBdev2", 00:28:54.412 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:54.412 "is_configured": false, 00:28:54.412 "data_offset": 0, 00:28:54.412 "data_size": 0 00:28:54.412 }, 00:28:54.412 { 00:28:54.412 "name": "BaseBdev3", 00:28:54.412 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:54.412 "is_configured": false, 00:28:54.412 "data_offset": 0, 00:28:54.412 "data_size": 0 00:28:54.412 }, 00:28:54.412 { 00:28:54.412 "name": "BaseBdev4", 00:28:54.412 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:54.412 "is_configured": false, 00:28:54.412 "data_offset": 0, 00:28:54.412 "data_size": 0 00:28:54.412 } 00:28:54.412 ] 00:28:54.412 }' 00:28:54.412 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:54.412 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:54.670 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:28:54.670 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:54.670 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:54.670 [2024-11-20 13:49:01.683769] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:28:54.670 [2024-11-20 13:49:01.683824] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:28:54.670 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:54.670 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:28:54.670 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:54.670 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:54.670 [2024-11-20 13:49:01.695770] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:28:54.670 [2024-11-20 13:49:01.695823] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:28:54.670 [2024-11-20 13:49:01.695838] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:28:54.670 [2024-11-20 13:49:01.695845] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:28:54.670 [2024-11-20 13:49:01.695851] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:28:54.670 [2024-11-20 13:49:01.695858] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:28:54.670 [2024-11-20 13:49:01.695863] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:28:54.670 [2024-11-20 13:49:01.695871] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:28:54.670 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:54.670 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:28:54.670 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:54.670 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:54.670 [2024-11-20 13:49:01.726505] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:54.927 BaseBdev1 00:28:54.927 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:54.927 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:28:54.927 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:28:54.927 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:54.927 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:54.928 [ 00:28:54.928 { 00:28:54.928 "name": "BaseBdev1", 00:28:54.928 "aliases": [ 00:28:54.928 "2857f008-ed3c-41ec-a638-45341deff93e" 00:28:54.928 ], 00:28:54.928 "product_name": "Malloc disk", 00:28:54.928 "block_size": 512, 00:28:54.928 "num_blocks": 65536, 00:28:54.928 "uuid": "2857f008-ed3c-41ec-a638-45341deff93e", 00:28:54.928 "assigned_rate_limits": { 00:28:54.928 "rw_ios_per_sec": 0, 00:28:54.928 "rw_mbytes_per_sec": 0, 00:28:54.928 "r_mbytes_per_sec": 0, 00:28:54.928 "w_mbytes_per_sec": 0 00:28:54.928 }, 00:28:54.928 "claimed": true, 00:28:54.928 "claim_type": "exclusive_write", 00:28:54.928 "zoned": false, 00:28:54.928 "supported_io_types": { 00:28:54.928 "read": true, 00:28:54.928 "write": true, 00:28:54.928 "unmap": true, 00:28:54.928 "flush": true, 00:28:54.928 "reset": true, 00:28:54.928 "nvme_admin": false, 00:28:54.928 "nvme_io": false, 00:28:54.928 "nvme_io_md": false, 00:28:54.928 "write_zeroes": true, 00:28:54.928 "zcopy": true, 00:28:54.928 "get_zone_info": false, 00:28:54.928 "zone_management": false, 00:28:54.928 "zone_append": false, 00:28:54.928 "compare": false, 00:28:54.928 "compare_and_write": false, 00:28:54.928 "abort": true, 00:28:54.928 "seek_hole": false, 00:28:54.928 "seek_data": false, 00:28:54.928 "copy": true, 00:28:54.928 "nvme_iov_md": false 00:28:54.928 }, 00:28:54.928 "memory_domains": [ 00:28:54.928 { 00:28:54.928 "dma_device_id": "system", 00:28:54.928 "dma_device_type": 1 00:28:54.928 }, 00:28:54.928 { 00:28:54.928 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:54.928 "dma_device_type": 2 00:28:54.928 } 00:28:54.928 ], 00:28:54.928 "driver_specific": {} 00:28:54.928 } 00:28:54.928 ] 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:54.928 "name": "Existed_Raid", 00:28:54.928 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:54.928 "strip_size_kb": 64, 00:28:54.928 "state": "configuring", 00:28:54.928 "raid_level": "raid0", 00:28:54.928 "superblock": false, 00:28:54.928 "num_base_bdevs": 4, 00:28:54.928 "num_base_bdevs_discovered": 1, 00:28:54.928 "num_base_bdevs_operational": 4, 00:28:54.928 "base_bdevs_list": [ 00:28:54.928 { 00:28:54.928 "name": "BaseBdev1", 00:28:54.928 "uuid": "2857f008-ed3c-41ec-a638-45341deff93e", 00:28:54.928 "is_configured": true, 00:28:54.928 "data_offset": 0, 00:28:54.928 "data_size": 65536 00:28:54.928 }, 00:28:54.928 { 00:28:54.928 "name": "BaseBdev2", 00:28:54.928 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:54.928 "is_configured": false, 00:28:54.928 "data_offset": 0, 00:28:54.928 "data_size": 0 00:28:54.928 }, 00:28:54.928 { 00:28:54.928 "name": "BaseBdev3", 00:28:54.928 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:54.928 "is_configured": false, 00:28:54.928 "data_offset": 0, 00:28:54.928 "data_size": 0 00:28:54.928 }, 00:28:54.928 { 00:28:54.928 "name": "BaseBdev4", 00:28:54.928 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:54.928 "is_configured": false, 00:28:54.928 "data_offset": 0, 00:28:54.928 "data_size": 0 00:28:54.928 } 00:28:54.928 ] 00:28:54.928 }' 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:54.928 13:49:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:55.186 [2024-11-20 13:49:02.114683] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:28:55.186 [2024-11-20 13:49:02.114911] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:55.186 [2024-11-20 13:49:02.122720] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:55.186 [2024-11-20 13:49:02.124783] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:28:55.186 [2024-11-20 13:49:02.124915] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:28:55.186 [2024-11-20 13:49:02.125000] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:28:55.186 [2024-11-20 13:49:02.125031] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:28:55.186 [2024-11-20 13:49:02.125078] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:28:55.186 [2024-11-20 13:49:02.125090] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:55.186 "name": "Existed_Raid", 00:28:55.186 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:55.186 "strip_size_kb": 64, 00:28:55.186 "state": "configuring", 00:28:55.186 "raid_level": "raid0", 00:28:55.186 "superblock": false, 00:28:55.186 "num_base_bdevs": 4, 00:28:55.186 "num_base_bdevs_discovered": 1, 00:28:55.186 "num_base_bdevs_operational": 4, 00:28:55.186 "base_bdevs_list": [ 00:28:55.186 { 00:28:55.186 "name": "BaseBdev1", 00:28:55.186 "uuid": "2857f008-ed3c-41ec-a638-45341deff93e", 00:28:55.186 "is_configured": true, 00:28:55.186 "data_offset": 0, 00:28:55.186 "data_size": 65536 00:28:55.186 }, 00:28:55.186 { 00:28:55.186 "name": "BaseBdev2", 00:28:55.186 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:55.186 "is_configured": false, 00:28:55.186 "data_offset": 0, 00:28:55.186 "data_size": 0 00:28:55.186 }, 00:28:55.186 { 00:28:55.186 "name": "BaseBdev3", 00:28:55.186 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:55.186 "is_configured": false, 00:28:55.186 "data_offset": 0, 00:28:55.186 "data_size": 0 00:28:55.186 }, 00:28:55.186 { 00:28:55.186 "name": "BaseBdev4", 00:28:55.186 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:55.186 "is_configured": false, 00:28:55.186 "data_offset": 0, 00:28:55.186 "data_size": 0 00:28:55.186 } 00:28:55.186 ] 00:28:55.186 }' 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:55.186 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:55.444 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:28:55.444 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.444 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:55.444 [2024-11-20 13:49:02.488164] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:28:55.444 BaseBdev2 00:28:55.444 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:55.444 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:28:55.444 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:28:55.444 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:55.444 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:28:55.444 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:55.444 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:55.444 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:55.444 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.444 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:55.702 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:55.702 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:28:55.702 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.702 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:55.702 [ 00:28:55.702 { 00:28:55.702 "name": "BaseBdev2", 00:28:55.702 "aliases": [ 00:28:55.702 "b7c4a059-00a0-4284-959e-4098c9220fff" 00:28:55.702 ], 00:28:55.702 "product_name": "Malloc disk", 00:28:55.702 "block_size": 512, 00:28:55.702 "num_blocks": 65536, 00:28:55.702 "uuid": "b7c4a059-00a0-4284-959e-4098c9220fff", 00:28:55.702 "assigned_rate_limits": { 00:28:55.702 "rw_ios_per_sec": 0, 00:28:55.702 "rw_mbytes_per_sec": 0, 00:28:55.702 "r_mbytes_per_sec": 0, 00:28:55.702 "w_mbytes_per_sec": 0 00:28:55.702 }, 00:28:55.702 "claimed": true, 00:28:55.702 "claim_type": "exclusive_write", 00:28:55.702 "zoned": false, 00:28:55.702 "supported_io_types": { 00:28:55.702 "read": true, 00:28:55.702 "write": true, 00:28:55.702 "unmap": true, 00:28:55.702 "flush": true, 00:28:55.702 "reset": true, 00:28:55.702 "nvme_admin": false, 00:28:55.702 "nvme_io": false, 00:28:55.702 "nvme_io_md": false, 00:28:55.702 "write_zeroes": true, 00:28:55.702 "zcopy": true, 00:28:55.702 "get_zone_info": false, 00:28:55.702 "zone_management": false, 00:28:55.702 "zone_append": false, 00:28:55.702 "compare": false, 00:28:55.702 "compare_and_write": false, 00:28:55.702 "abort": true, 00:28:55.702 "seek_hole": false, 00:28:55.702 "seek_data": false, 00:28:55.702 "copy": true, 00:28:55.703 "nvme_iov_md": false 00:28:55.703 }, 00:28:55.703 "memory_domains": [ 00:28:55.703 { 00:28:55.703 "dma_device_id": "system", 00:28:55.703 "dma_device_type": 1 00:28:55.703 }, 00:28:55.703 { 00:28:55.703 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:55.703 "dma_device_type": 2 00:28:55.703 } 00:28:55.703 ], 00:28:55.703 "driver_specific": {} 00:28:55.703 } 00:28:55.703 ] 00:28:55.703 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:55.703 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:28:55.703 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:28:55.703 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:28:55.703 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:28:55.703 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:55.703 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:55.703 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:28:55.703 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:55.703 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:28:55.703 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:55.703 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:55.703 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:55.703 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:55.703 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:55.703 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:55.703 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.703 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:55.703 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:55.703 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:55.703 "name": "Existed_Raid", 00:28:55.703 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:55.703 "strip_size_kb": 64, 00:28:55.703 "state": "configuring", 00:28:55.703 "raid_level": "raid0", 00:28:55.703 "superblock": false, 00:28:55.703 "num_base_bdevs": 4, 00:28:55.703 "num_base_bdevs_discovered": 2, 00:28:55.703 "num_base_bdevs_operational": 4, 00:28:55.703 "base_bdevs_list": [ 00:28:55.703 { 00:28:55.703 "name": "BaseBdev1", 00:28:55.703 "uuid": "2857f008-ed3c-41ec-a638-45341deff93e", 00:28:55.703 "is_configured": true, 00:28:55.703 "data_offset": 0, 00:28:55.703 "data_size": 65536 00:28:55.703 }, 00:28:55.703 { 00:28:55.703 "name": "BaseBdev2", 00:28:55.703 "uuid": "b7c4a059-00a0-4284-959e-4098c9220fff", 00:28:55.703 "is_configured": true, 00:28:55.703 "data_offset": 0, 00:28:55.703 "data_size": 65536 00:28:55.703 }, 00:28:55.703 { 00:28:55.703 "name": "BaseBdev3", 00:28:55.703 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:55.703 "is_configured": false, 00:28:55.703 "data_offset": 0, 00:28:55.703 "data_size": 0 00:28:55.703 }, 00:28:55.703 { 00:28:55.703 "name": "BaseBdev4", 00:28:55.703 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:55.703 "is_configured": false, 00:28:55.703 "data_offset": 0, 00:28:55.703 "data_size": 0 00:28:55.703 } 00:28:55.703 ] 00:28:55.703 }' 00:28:55.703 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:55.703 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:55.961 [2024-11-20 13:49:02.966932] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:28:55.961 BaseBdev3 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:55.961 [ 00:28:55.961 { 00:28:55.961 "name": "BaseBdev3", 00:28:55.961 "aliases": [ 00:28:55.961 "d5ee6ca9-5fbb-424c-bc58-9d439ee6d243" 00:28:55.961 ], 00:28:55.961 "product_name": "Malloc disk", 00:28:55.961 "block_size": 512, 00:28:55.961 "num_blocks": 65536, 00:28:55.961 "uuid": "d5ee6ca9-5fbb-424c-bc58-9d439ee6d243", 00:28:55.961 "assigned_rate_limits": { 00:28:55.961 "rw_ios_per_sec": 0, 00:28:55.961 "rw_mbytes_per_sec": 0, 00:28:55.961 "r_mbytes_per_sec": 0, 00:28:55.961 "w_mbytes_per_sec": 0 00:28:55.961 }, 00:28:55.961 "claimed": true, 00:28:55.961 "claim_type": "exclusive_write", 00:28:55.961 "zoned": false, 00:28:55.961 "supported_io_types": { 00:28:55.961 "read": true, 00:28:55.961 "write": true, 00:28:55.961 "unmap": true, 00:28:55.961 "flush": true, 00:28:55.961 "reset": true, 00:28:55.961 "nvme_admin": false, 00:28:55.961 "nvme_io": false, 00:28:55.961 "nvme_io_md": false, 00:28:55.961 "write_zeroes": true, 00:28:55.961 "zcopy": true, 00:28:55.961 "get_zone_info": false, 00:28:55.961 "zone_management": false, 00:28:55.961 "zone_append": false, 00:28:55.961 "compare": false, 00:28:55.961 "compare_and_write": false, 00:28:55.961 "abort": true, 00:28:55.961 "seek_hole": false, 00:28:55.961 "seek_data": false, 00:28:55.961 "copy": true, 00:28:55.961 "nvme_iov_md": false 00:28:55.961 }, 00:28:55.961 "memory_domains": [ 00:28:55.961 { 00:28:55.961 "dma_device_id": "system", 00:28:55.961 "dma_device_type": 1 00:28:55.961 }, 00:28:55.961 { 00:28:55.961 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:55.961 "dma_device_type": 2 00:28:55.961 } 00:28:55.961 ], 00:28:55.961 "driver_specific": {} 00:28:55.961 } 00:28:55.961 ] 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:55.961 13:49:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:55.961 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:56.220 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:56.220 "name": "Existed_Raid", 00:28:56.220 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:56.220 "strip_size_kb": 64, 00:28:56.220 "state": "configuring", 00:28:56.220 "raid_level": "raid0", 00:28:56.220 "superblock": false, 00:28:56.220 "num_base_bdevs": 4, 00:28:56.220 "num_base_bdevs_discovered": 3, 00:28:56.220 "num_base_bdevs_operational": 4, 00:28:56.220 "base_bdevs_list": [ 00:28:56.220 { 00:28:56.220 "name": "BaseBdev1", 00:28:56.220 "uuid": "2857f008-ed3c-41ec-a638-45341deff93e", 00:28:56.220 "is_configured": true, 00:28:56.220 "data_offset": 0, 00:28:56.220 "data_size": 65536 00:28:56.220 }, 00:28:56.220 { 00:28:56.220 "name": "BaseBdev2", 00:28:56.220 "uuid": "b7c4a059-00a0-4284-959e-4098c9220fff", 00:28:56.220 "is_configured": true, 00:28:56.220 "data_offset": 0, 00:28:56.220 "data_size": 65536 00:28:56.220 }, 00:28:56.220 { 00:28:56.220 "name": "BaseBdev3", 00:28:56.220 "uuid": "d5ee6ca9-5fbb-424c-bc58-9d439ee6d243", 00:28:56.220 "is_configured": true, 00:28:56.220 "data_offset": 0, 00:28:56.220 "data_size": 65536 00:28:56.220 }, 00:28:56.220 { 00:28:56.220 "name": "BaseBdev4", 00:28:56.220 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:56.220 "is_configured": false, 00:28:56.220 "data_offset": 0, 00:28:56.220 "data_size": 0 00:28:56.220 } 00:28:56.220 ] 00:28:56.220 }' 00:28:56.220 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:56.220 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:56.505 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:28:56.505 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:56.505 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:56.505 [2024-11-20 13:49:03.372339] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:28:56.505 [2024-11-20 13:49:03.372654] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:28:56.505 [2024-11-20 13:49:03.372679] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:28:56.505 [2024-11-20 13:49:03.373097] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:28:56.505 [2024-11-20 13:49:03.373311] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:28:56.505 [2024-11-20 13:49:03.373325] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:28:56.505 [2024-11-20 13:49:03.373678] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:56.505 BaseBdev4 00:28:56.505 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:56.505 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:28:56.505 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:28:56.505 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:56.505 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:28:56.505 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:56.505 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:56.505 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:56.505 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:56.505 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:56.505 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:56.505 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:28:56.505 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:56.505 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:56.505 [ 00:28:56.505 { 00:28:56.505 "name": "BaseBdev4", 00:28:56.505 "aliases": [ 00:28:56.505 "bf202031-6f27-45a7-b472-3f50e6e8ecbf" 00:28:56.505 ], 00:28:56.505 "product_name": "Malloc disk", 00:28:56.505 "block_size": 512, 00:28:56.505 "num_blocks": 65536, 00:28:56.505 "uuid": "bf202031-6f27-45a7-b472-3f50e6e8ecbf", 00:28:56.505 "assigned_rate_limits": { 00:28:56.505 "rw_ios_per_sec": 0, 00:28:56.505 "rw_mbytes_per_sec": 0, 00:28:56.505 "r_mbytes_per_sec": 0, 00:28:56.505 "w_mbytes_per_sec": 0 00:28:56.505 }, 00:28:56.505 "claimed": true, 00:28:56.505 "claim_type": "exclusive_write", 00:28:56.505 "zoned": false, 00:28:56.505 "supported_io_types": { 00:28:56.505 "read": true, 00:28:56.506 "write": true, 00:28:56.506 "unmap": true, 00:28:56.506 "flush": true, 00:28:56.506 "reset": true, 00:28:56.506 "nvme_admin": false, 00:28:56.506 "nvme_io": false, 00:28:56.506 "nvme_io_md": false, 00:28:56.506 "write_zeroes": true, 00:28:56.506 "zcopy": true, 00:28:56.506 "get_zone_info": false, 00:28:56.506 "zone_management": false, 00:28:56.506 "zone_append": false, 00:28:56.506 "compare": false, 00:28:56.506 "compare_and_write": false, 00:28:56.506 "abort": true, 00:28:56.506 "seek_hole": false, 00:28:56.506 "seek_data": false, 00:28:56.506 "copy": true, 00:28:56.506 "nvme_iov_md": false 00:28:56.506 }, 00:28:56.506 "memory_domains": [ 00:28:56.506 { 00:28:56.506 "dma_device_id": "system", 00:28:56.506 "dma_device_type": 1 00:28:56.506 }, 00:28:56.506 { 00:28:56.506 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:56.506 "dma_device_type": 2 00:28:56.506 } 00:28:56.506 ], 00:28:56.506 "driver_specific": {} 00:28:56.506 } 00:28:56.506 ] 00:28:56.506 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:56.506 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:28:56.506 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:28:56.506 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:28:56.506 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:28:56.506 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:56.506 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:56.506 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:28:56.506 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:56.506 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:28:56.506 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:56.506 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:56.506 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:56.506 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:56.506 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:56.506 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:56.506 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:56.506 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:56.506 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:56.506 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:56.506 "name": "Existed_Raid", 00:28:56.506 "uuid": "6bf776de-82cc-40da-b1a0-c1cd3106fa84", 00:28:56.506 "strip_size_kb": 64, 00:28:56.506 "state": "online", 00:28:56.506 "raid_level": "raid0", 00:28:56.506 "superblock": false, 00:28:56.506 "num_base_bdevs": 4, 00:28:56.506 "num_base_bdevs_discovered": 4, 00:28:56.506 "num_base_bdevs_operational": 4, 00:28:56.506 "base_bdevs_list": [ 00:28:56.506 { 00:28:56.506 "name": "BaseBdev1", 00:28:56.506 "uuid": "2857f008-ed3c-41ec-a638-45341deff93e", 00:28:56.506 "is_configured": true, 00:28:56.506 "data_offset": 0, 00:28:56.506 "data_size": 65536 00:28:56.506 }, 00:28:56.506 { 00:28:56.506 "name": "BaseBdev2", 00:28:56.506 "uuid": "b7c4a059-00a0-4284-959e-4098c9220fff", 00:28:56.506 "is_configured": true, 00:28:56.506 "data_offset": 0, 00:28:56.506 "data_size": 65536 00:28:56.506 }, 00:28:56.506 { 00:28:56.506 "name": "BaseBdev3", 00:28:56.506 "uuid": "d5ee6ca9-5fbb-424c-bc58-9d439ee6d243", 00:28:56.506 "is_configured": true, 00:28:56.506 "data_offset": 0, 00:28:56.506 "data_size": 65536 00:28:56.506 }, 00:28:56.506 { 00:28:56.506 "name": "BaseBdev4", 00:28:56.506 "uuid": "bf202031-6f27-45a7-b472-3f50e6e8ecbf", 00:28:56.506 "is_configured": true, 00:28:56.506 "data_offset": 0, 00:28:56.506 "data_size": 65536 00:28:56.506 } 00:28:56.506 ] 00:28:56.506 }' 00:28:56.506 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:56.506 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:56.764 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:28:56.764 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:28:56.764 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:28:56.764 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:28:56.764 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:28:56.764 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:28:56.764 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:28:56.764 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:28:56.764 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:56.764 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:56.764 [2024-11-20 13:49:03.752764] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:56.764 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:56.764 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:28:56.764 "name": "Existed_Raid", 00:28:56.764 "aliases": [ 00:28:56.764 "6bf776de-82cc-40da-b1a0-c1cd3106fa84" 00:28:56.764 ], 00:28:56.764 "product_name": "Raid Volume", 00:28:56.764 "block_size": 512, 00:28:56.764 "num_blocks": 262144, 00:28:56.764 "uuid": "6bf776de-82cc-40da-b1a0-c1cd3106fa84", 00:28:56.764 "assigned_rate_limits": { 00:28:56.764 "rw_ios_per_sec": 0, 00:28:56.764 "rw_mbytes_per_sec": 0, 00:28:56.764 "r_mbytes_per_sec": 0, 00:28:56.764 "w_mbytes_per_sec": 0 00:28:56.764 }, 00:28:56.764 "claimed": false, 00:28:56.764 "zoned": false, 00:28:56.764 "supported_io_types": { 00:28:56.764 "read": true, 00:28:56.764 "write": true, 00:28:56.764 "unmap": true, 00:28:56.764 "flush": true, 00:28:56.764 "reset": true, 00:28:56.764 "nvme_admin": false, 00:28:56.764 "nvme_io": false, 00:28:56.764 "nvme_io_md": false, 00:28:56.764 "write_zeroes": true, 00:28:56.764 "zcopy": false, 00:28:56.764 "get_zone_info": false, 00:28:56.764 "zone_management": false, 00:28:56.764 "zone_append": false, 00:28:56.765 "compare": false, 00:28:56.765 "compare_and_write": false, 00:28:56.765 "abort": false, 00:28:56.765 "seek_hole": false, 00:28:56.765 "seek_data": false, 00:28:56.765 "copy": false, 00:28:56.765 "nvme_iov_md": false 00:28:56.765 }, 00:28:56.765 "memory_domains": [ 00:28:56.765 { 00:28:56.765 "dma_device_id": "system", 00:28:56.765 "dma_device_type": 1 00:28:56.765 }, 00:28:56.765 { 00:28:56.765 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:56.765 "dma_device_type": 2 00:28:56.765 }, 00:28:56.765 { 00:28:56.765 "dma_device_id": "system", 00:28:56.765 "dma_device_type": 1 00:28:56.765 }, 00:28:56.765 { 00:28:56.765 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:56.765 "dma_device_type": 2 00:28:56.765 }, 00:28:56.765 { 00:28:56.765 "dma_device_id": "system", 00:28:56.765 "dma_device_type": 1 00:28:56.765 }, 00:28:56.765 { 00:28:56.765 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:56.765 "dma_device_type": 2 00:28:56.765 }, 00:28:56.765 { 00:28:56.765 "dma_device_id": "system", 00:28:56.765 "dma_device_type": 1 00:28:56.765 }, 00:28:56.765 { 00:28:56.765 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:56.765 "dma_device_type": 2 00:28:56.765 } 00:28:56.765 ], 00:28:56.765 "driver_specific": { 00:28:56.765 "raid": { 00:28:56.765 "uuid": "6bf776de-82cc-40da-b1a0-c1cd3106fa84", 00:28:56.765 "strip_size_kb": 64, 00:28:56.765 "state": "online", 00:28:56.765 "raid_level": "raid0", 00:28:56.765 "superblock": false, 00:28:56.765 "num_base_bdevs": 4, 00:28:56.765 "num_base_bdevs_discovered": 4, 00:28:56.765 "num_base_bdevs_operational": 4, 00:28:56.765 "base_bdevs_list": [ 00:28:56.765 { 00:28:56.765 "name": "BaseBdev1", 00:28:56.765 "uuid": "2857f008-ed3c-41ec-a638-45341deff93e", 00:28:56.765 "is_configured": true, 00:28:56.765 "data_offset": 0, 00:28:56.765 "data_size": 65536 00:28:56.765 }, 00:28:56.765 { 00:28:56.765 "name": "BaseBdev2", 00:28:56.765 "uuid": "b7c4a059-00a0-4284-959e-4098c9220fff", 00:28:56.765 "is_configured": true, 00:28:56.765 "data_offset": 0, 00:28:56.765 "data_size": 65536 00:28:56.765 }, 00:28:56.765 { 00:28:56.765 "name": "BaseBdev3", 00:28:56.765 "uuid": "d5ee6ca9-5fbb-424c-bc58-9d439ee6d243", 00:28:56.765 "is_configured": true, 00:28:56.765 "data_offset": 0, 00:28:56.765 "data_size": 65536 00:28:56.765 }, 00:28:56.765 { 00:28:56.765 "name": "BaseBdev4", 00:28:56.765 "uuid": "bf202031-6f27-45a7-b472-3f50e6e8ecbf", 00:28:56.765 "is_configured": true, 00:28:56.765 "data_offset": 0, 00:28:56.765 "data_size": 65536 00:28:56.765 } 00:28:56.765 ] 00:28:56.765 } 00:28:56.765 } 00:28:56.765 }' 00:28:56.765 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:28:56.765 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:28:56.765 BaseBdev2 00:28:56.765 BaseBdev3 00:28:56.765 BaseBdev4' 00:28:56.765 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.023 13:49:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.023 [2024-11-20 13:49:03.984530] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:28:57.023 [2024-11-20 13:49:03.984710] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:28:57.023 [2024-11-20 13:49:03.984772] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:57.023 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.023 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:28:57.023 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:28:57.023 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:28:57.023 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:28:57.023 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:28:57.023 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 3 00:28:57.023 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:57.023 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:28:57.023 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:28:57.023 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:57.023 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:57.023 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:57.023 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:57.023 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:57.023 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:57.023 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:57.023 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:57.023 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.023 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.023 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.023 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:57.023 "name": "Existed_Raid", 00:28:57.023 "uuid": "6bf776de-82cc-40da-b1a0-c1cd3106fa84", 00:28:57.023 "strip_size_kb": 64, 00:28:57.024 "state": "offline", 00:28:57.024 "raid_level": "raid0", 00:28:57.024 "superblock": false, 00:28:57.024 "num_base_bdevs": 4, 00:28:57.024 "num_base_bdevs_discovered": 3, 00:28:57.024 "num_base_bdevs_operational": 3, 00:28:57.024 "base_bdevs_list": [ 00:28:57.024 { 00:28:57.024 "name": null, 00:28:57.024 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:57.024 "is_configured": false, 00:28:57.024 "data_offset": 0, 00:28:57.024 "data_size": 65536 00:28:57.024 }, 00:28:57.024 { 00:28:57.024 "name": "BaseBdev2", 00:28:57.024 "uuid": "b7c4a059-00a0-4284-959e-4098c9220fff", 00:28:57.024 "is_configured": true, 00:28:57.024 "data_offset": 0, 00:28:57.024 "data_size": 65536 00:28:57.024 }, 00:28:57.024 { 00:28:57.024 "name": "BaseBdev3", 00:28:57.024 "uuid": "d5ee6ca9-5fbb-424c-bc58-9d439ee6d243", 00:28:57.024 "is_configured": true, 00:28:57.024 "data_offset": 0, 00:28:57.024 "data_size": 65536 00:28:57.024 }, 00:28:57.024 { 00:28:57.024 "name": "BaseBdev4", 00:28:57.024 "uuid": "bf202031-6f27-45a7-b472-3f50e6e8ecbf", 00:28:57.024 "is_configured": true, 00:28:57.024 "data_offset": 0, 00:28:57.024 "data_size": 65536 00:28:57.024 } 00:28:57.024 ] 00:28:57.024 }' 00:28:57.024 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:57.024 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.589 [2024-11-20 13:49:04.414997] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.589 [2024-11-20 13:49:04.505518] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.589 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.589 [2024-11-20 13:49:04.611374] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:28:57.589 [2024-11-20 13:49:04.611434] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:28:57.848 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.848 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:28:57.848 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:28:57.848 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.849 BaseBdev2 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.849 [ 00:28:57.849 { 00:28:57.849 "name": "BaseBdev2", 00:28:57.849 "aliases": [ 00:28:57.849 "989e0202-545d-4fc7-848f-1f98f8fdd9b6" 00:28:57.849 ], 00:28:57.849 "product_name": "Malloc disk", 00:28:57.849 "block_size": 512, 00:28:57.849 "num_blocks": 65536, 00:28:57.849 "uuid": "989e0202-545d-4fc7-848f-1f98f8fdd9b6", 00:28:57.849 "assigned_rate_limits": { 00:28:57.849 "rw_ios_per_sec": 0, 00:28:57.849 "rw_mbytes_per_sec": 0, 00:28:57.849 "r_mbytes_per_sec": 0, 00:28:57.849 "w_mbytes_per_sec": 0 00:28:57.849 }, 00:28:57.849 "claimed": false, 00:28:57.849 "zoned": false, 00:28:57.849 "supported_io_types": { 00:28:57.849 "read": true, 00:28:57.849 "write": true, 00:28:57.849 "unmap": true, 00:28:57.849 "flush": true, 00:28:57.849 "reset": true, 00:28:57.849 "nvme_admin": false, 00:28:57.849 "nvme_io": false, 00:28:57.849 "nvme_io_md": false, 00:28:57.849 "write_zeroes": true, 00:28:57.849 "zcopy": true, 00:28:57.849 "get_zone_info": false, 00:28:57.849 "zone_management": false, 00:28:57.849 "zone_append": false, 00:28:57.849 "compare": false, 00:28:57.849 "compare_and_write": false, 00:28:57.849 "abort": true, 00:28:57.849 "seek_hole": false, 00:28:57.849 "seek_data": false, 00:28:57.849 "copy": true, 00:28:57.849 "nvme_iov_md": false 00:28:57.849 }, 00:28:57.849 "memory_domains": [ 00:28:57.849 { 00:28:57.849 "dma_device_id": "system", 00:28:57.849 "dma_device_type": 1 00:28:57.849 }, 00:28:57.849 { 00:28:57.849 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:57.849 "dma_device_type": 2 00:28:57.849 } 00:28:57.849 ], 00:28:57.849 "driver_specific": {} 00:28:57.849 } 00:28:57.849 ] 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.849 BaseBdev3 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.849 [ 00:28:57.849 { 00:28:57.849 "name": "BaseBdev3", 00:28:57.849 "aliases": [ 00:28:57.849 "52a30748-1767-45e5-9109-cabee6bdc9a8" 00:28:57.849 ], 00:28:57.849 "product_name": "Malloc disk", 00:28:57.849 "block_size": 512, 00:28:57.849 "num_blocks": 65536, 00:28:57.849 "uuid": "52a30748-1767-45e5-9109-cabee6bdc9a8", 00:28:57.849 "assigned_rate_limits": { 00:28:57.849 "rw_ios_per_sec": 0, 00:28:57.849 "rw_mbytes_per_sec": 0, 00:28:57.849 "r_mbytes_per_sec": 0, 00:28:57.849 "w_mbytes_per_sec": 0 00:28:57.849 }, 00:28:57.849 "claimed": false, 00:28:57.849 "zoned": false, 00:28:57.849 "supported_io_types": { 00:28:57.849 "read": true, 00:28:57.849 "write": true, 00:28:57.849 "unmap": true, 00:28:57.849 "flush": true, 00:28:57.849 "reset": true, 00:28:57.849 "nvme_admin": false, 00:28:57.849 "nvme_io": false, 00:28:57.849 "nvme_io_md": false, 00:28:57.849 "write_zeroes": true, 00:28:57.849 "zcopy": true, 00:28:57.849 "get_zone_info": false, 00:28:57.849 "zone_management": false, 00:28:57.849 "zone_append": false, 00:28:57.849 "compare": false, 00:28:57.849 "compare_and_write": false, 00:28:57.849 "abort": true, 00:28:57.849 "seek_hole": false, 00:28:57.849 "seek_data": false, 00:28:57.849 "copy": true, 00:28:57.849 "nvme_iov_md": false 00:28:57.849 }, 00:28:57.849 "memory_domains": [ 00:28:57.849 { 00:28:57.849 "dma_device_id": "system", 00:28:57.849 "dma_device_type": 1 00:28:57.849 }, 00:28:57.849 { 00:28:57.849 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:57.849 "dma_device_type": 2 00:28:57.849 } 00:28:57.849 ], 00:28:57.849 "driver_specific": {} 00:28:57.849 } 00:28:57.849 ] 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.849 BaseBdev4 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.849 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.849 [ 00:28:57.849 { 00:28:57.849 "name": "BaseBdev4", 00:28:57.849 "aliases": [ 00:28:57.849 "648c9b30-6d19-47e6-a202-109991a1150a" 00:28:57.849 ], 00:28:57.849 "product_name": "Malloc disk", 00:28:57.850 "block_size": 512, 00:28:57.850 "num_blocks": 65536, 00:28:57.850 "uuid": "648c9b30-6d19-47e6-a202-109991a1150a", 00:28:57.850 "assigned_rate_limits": { 00:28:57.850 "rw_ios_per_sec": 0, 00:28:57.850 "rw_mbytes_per_sec": 0, 00:28:57.850 "r_mbytes_per_sec": 0, 00:28:57.850 "w_mbytes_per_sec": 0 00:28:57.850 }, 00:28:57.850 "claimed": false, 00:28:57.850 "zoned": false, 00:28:57.850 "supported_io_types": { 00:28:57.850 "read": true, 00:28:57.850 "write": true, 00:28:57.850 "unmap": true, 00:28:57.850 "flush": true, 00:28:57.850 "reset": true, 00:28:57.850 "nvme_admin": false, 00:28:57.850 "nvme_io": false, 00:28:57.850 "nvme_io_md": false, 00:28:57.850 "write_zeroes": true, 00:28:57.850 "zcopy": true, 00:28:57.850 "get_zone_info": false, 00:28:57.850 "zone_management": false, 00:28:57.850 "zone_append": false, 00:28:57.850 "compare": false, 00:28:57.850 "compare_and_write": false, 00:28:57.850 "abort": true, 00:28:57.850 "seek_hole": false, 00:28:57.850 "seek_data": false, 00:28:57.850 "copy": true, 00:28:57.850 "nvme_iov_md": false 00:28:57.850 }, 00:28:57.850 "memory_domains": [ 00:28:57.850 { 00:28:57.850 "dma_device_id": "system", 00:28:57.850 "dma_device_type": 1 00:28:57.850 }, 00:28:57.850 { 00:28:57.850 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:57.850 "dma_device_type": 2 00:28:57.850 } 00:28:57.850 ], 00:28:57.850 "driver_specific": {} 00:28:57.850 } 00:28:57.850 ] 00:28:57.850 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.850 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:28:57.850 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:28:57.850 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:28:57.850 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:28:57.850 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.850 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.850 [2024-11-20 13:49:04.865321] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:28:57.850 [2024-11-20 13:49:04.865504] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:28:57.850 [2024-11-20 13:49:04.865570] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:28:57.850 [2024-11-20 13:49:04.867281] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:28:57.850 [2024-11-20 13:49:04.867402] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:28:57.850 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:57.850 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:28:57.850 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:57.850 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:57.850 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:28:57.850 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:57.850 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:28:57.850 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:57.850 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:57.850 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:57.850 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:57.850 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:57.850 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:57.850 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:57.850 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.850 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:58.108 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:58.108 "name": "Existed_Raid", 00:28:58.108 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:58.108 "strip_size_kb": 64, 00:28:58.108 "state": "configuring", 00:28:58.108 "raid_level": "raid0", 00:28:58.108 "superblock": false, 00:28:58.108 "num_base_bdevs": 4, 00:28:58.108 "num_base_bdevs_discovered": 3, 00:28:58.108 "num_base_bdevs_operational": 4, 00:28:58.108 "base_bdevs_list": [ 00:28:58.108 { 00:28:58.108 "name": "BaseBdev1", 00:28:58.108 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:58.108 "is_configured": false, 00:28:58.108 "data_offset": 0, 00:28:58.108 "data_size": 0 00:28:58.108 }, 00:28:58.108 { 00:28:58.108 "name": "BaseBdev2", 00:28:58.108 "uuid": "989e0202-545d-4fc7-848f-1f98f8fdd9b6", 00:28:58.108 "is_configured": true, 00:28:58.108 "data_offset": 0, 00:28:58.108 "data_size": 65536 00:28:58.108 }, 00:28:58.108 { 00:28:58.108 "name": "BaseBdev3", 00:28:58.108 "uuid": "52a30748-1767-45e5-9109-cabee6bdc9a8", 00:28:58.108 "is_configured": true, 00:28:58.108 "data_offset": 0, 00:28:58.108 "data_size": 65536 00:28:58.108 }, 00:28:58.108 { 00:28:58.108 "name": "BaseBdev4", 00:28:58.108 "uuid": "648c9b30-6d19-47e6-a202-109991a1150a", 00:28:58.108 "is_configured": true, 00:28:58.108 "data_offset": 0, 00:28:58.108 "data_size": 65536 00:28:58.108 } 00:28:58.108 ] 00:28:58.108 }' 00:28:58.108 13:49:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:58.108 13:49:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:58.367 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:28:58.367 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:58.367 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:58.367 [2024-11-20 13:49:05.169425] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:28:58.367 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:58.367 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:28:58.367 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:58.367 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:58.367 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:28:58.367 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:58.367 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:28:58.367 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:58.367 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:58.367 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:58.367 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:58.367 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:58.367 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:58.367 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:58.367 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:58.367 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:58.367 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:58.367 "name": "Existed_Raid", 00:28:58.367 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:58.367 "strip_size_kb": 64, 00:28:58.367 "state": "configuring", 00:28:58.367 "raid_level": "raid0", 00:28:58.367 "superblock": false, 00:28:58.367 "num_base_bdevs": 4, 00:28:58.367 "num_base_bdevs_discovered": 2, 00:28:58.367 "num_base_bdevs_operational": 4, 00:28:58.367 "base_bdevs_list": [ 00:28:58.367 { 00:28:58.367 "name": "BaseBdev1", 00:28:58.367 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:58.367 "is_configured": false, 00:28:58.367 "data_offset": 0, 00:28:58.367 "data_size": 0 00:28:58.367 }, 00:28:58.367 { 00:28:58.367 "name": null, 00:28:58.367 "uuid": "989e0202-545d-4fc7-848f-1f98f8fdd9b6", 00:28:58.367 "is_configured": false, 00:28:58.367 "data_offset": 0, 00:28:58.367 "data_size": 65536 00:28:58.367 }, 00:28:58.367 { 00:28:58.367 "name": "BaseBdev3", 00:28:58.367 "uuid": "52a30748-1767-45e5-9109-cabee6bdc9a8", 00:28:58.367 "is_configured": true, 00:28:58.367 "data_offset": 0, 00:28:58.367 "data_size": 65536 00:28:58.367 }, 00:28:58.367 { 00:28:58.367 "name": "BaseBdev4", 00:28:58.367 "uuid": "648c9b30-6d19-47e6-a202-109991a1150a", 00:28:58.367 "is_configured": true, 00:28:58.367 "data_offset": 0, 00:28:58.367 "data_size": 65536 00:28:58.367 } 00:28:58.367 ] 00:28:58.367 }' 00:28:58.367 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:58.367 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:58.627 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:58.627 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:58.627 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:58.627 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:28:58.627 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:58.627 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:28:58.627 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:28:58.627 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:58.627 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:58.627 [2024-11-20 13:49:05.550387] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:58.627 BaseBdev1 00:28:58.627 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:58.627 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:28:58.627 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:28:58.627 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:28:58.627 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:28:58.627 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:28:58.627 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:28:58.627 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:28:58.627 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:58.627 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:58.627 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:58.627 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:28:58.627 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:58.627 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:58.627 [ 00:28:58.627 { 00:28:58.627 "name": "BaseBdev1", 00:28:58.627 "aliases": [ 00:28:58.627 "24d68baa-7898-45c7-9b41-8957c641fa4b" 00:28:58.627 ], 00:28:58.627 "product_name": "Malloc disk", 00:28:58.627 "block_size": 512, 00:28:58.627 "num_blocks": 65536, 00:28:58.627 "uuid": "24d68baa-7898-45c7-9b41-8957c641fa4b", 00:28:58.627 "assigned_rate_limits": { 00:28:58.627 "rw_ios_per_sec": 0, 00:28:58.627 "rw_mbytes_per_sec": 0, 00:28:58.627 "r_mbytes_per_sec": 0, 00:28:58.627 "w_mbytes_per_sec": 0 00:28:58.627 }, 00:28:58.627 "claimed": true, 00:28:58.627 "claim_type": "exclusive_write", 00:28:58.627 "zoned": false, 00:28:58.627 "supported_io_types": { 00:28:58.627 "read": true, 00:28:58.627 "write": true, 00:28:58.627 "unmap": true, 00:28:58.627 "flush": true, 00:28:58.627 "reset": true, 00:28:58.627 "nvme_admin": false, 00:28:58.627 "nvme_io": false, 00:28:58.627 "nvme_io_md": false, 00:28:58.627 "write_zeroes": true, 00:28:58.627 "zcopy": true, 00:28:58.627 "get_zone_info": false, 00:28:58.627 "zone_management": false, 00:28:58.627 "zone_append": false, 00:28:58.627 "compare": false, 00:28:58.627 "compare_and_write": false, 00:28:58.627 "abort": true, 00:28:58.627 "seek_hole": false, 00:28:58.627 "seek_data": false, 00:28:58.627 "copy": true, 00:28:58.627 "nvme_iov_md": false 00:28:58.627 }, 00:28:58.628 "memory_domains": [ 00:28:58.628 { 00:28:58.628 "dma_device_id": "system", 00:28:58.628 "dma_device_type": 1 00:28:58.628 }, 00:28:58.628 { 00:28:58.628 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:58.628 "dma_device_type": 2 00:28:58.628 } 00:28:58.628 ], 00:28:58.628 "driver_specific": {} 00:28:58.628 } 00:28:58.628 ] 00:28:58.628 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:58.628 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:28:58.628 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:28:58.628 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:58.628 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:58.628 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:28:58.628 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:58.628 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:28:58.628 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:58.628 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:58.628 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:58.628 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:58.628 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:58.628 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:58.628 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:58.628 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:58.628 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:58.628 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:58.628 "name": "Existed_Raid", 00:28:58.628 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:58.628 "strip_size_kb": 64, 00:28:58.628 "state": "configuring", 00:28:58.628 "raid_level": "raid0", 00:28:58.628 "superblock": false, 00:28:58.628 "num_base_bdevs": 4, 00:28:58.628 "num_base_bdevs_discovered": 3, 00:28:58.628 "num_base_bdevs_operational": 4, 00:28:58.628 "base_bdevs_list": [ 00:28:58.628 { 00:28:58.628 "name": "BaseBdev1", 00:28:58.628 "uuid": "24d68baa-7898-45c7-9b41-8957c641fa4b", 00:28:58.628 "is_configured": true, 00:28:58.628 "data_offset": 0, 00:28:58.628 "data_size": 65536 00:28:58.628 }, 00:28:58.628 { 00:28:58.628 "name": null, 00:28:58.628 "uuid": "989e0202-545d-4fc7-848f-1f98f8fdd9b6", 00:28:58.628 "is_configured": false, 00:28:58.628 "data_offset": 0, 00:28:58.628 "data_size": 65536 00:28:58.628 }, 00:28:58.628 { 00:28:58.628 "name": "BaseBdev3", 00:28:58.628 "uuid": "52a30748-1767-45e5-9109-cabee6bdc9a8", 00:28:58.628 "is_configured": true, 00:28:58.628 "data_offset": 0, 00:28:58.628 "data_size": 65536 00:28:58.628 }, 00:28:58.628 { 00:28:58.628 "name": "BaseBdev4", 00:28:58.628 "uuid": "648c9b30-6d19-47e6-a202-109991a1150a", 00:28:58.628 "is_configured": true, 00:28:58.628 "data_offset": 0, 00:28:58.628 "data_size": 65536 00:28:58.628 } 00:28:58.628 ] 00:28:58.628 }' 00:28:58.628 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:58.628 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:58.885 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:58.885 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:28:58.885 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:58.885 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:58.885 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:58.885 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:28:58.885 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:28:58.885 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:58.885 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:59.144 [2024-11-20 13:49:05.942574] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:28:59.144 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:59.144 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:28:59.144 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:59.144 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:59.144 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:28:59.144 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:59.144 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:28:59.144 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:59.144 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:59.144 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:59.144 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:59.144 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:59.144 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:59.144 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:59.144 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:59.144 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:59.144 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:59.144 "name": "Existed_Raid", 00:28:59.144 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:59.144 "strip_size_kb": 64, 00:28:59.144 "state": "configuring", 00:28:59.144 "raid_level": "raid0", 00:28:59.144 "superblock": false, 00:28:59.144 "num_base_bdevs": 4, 00:28:59.144 "num_base_bdevs_discovered": 2, 00:28:59.144 "num_base_bdevs_operational": 4, 00:28:59.144 "base_bdevs_list": [ 00:28:59.144 { 00:28:59.144 "name": "BaseBdev1", 00:28:59.144 "uuid": "24d68baa-7898-45c7-9b41-8957c641fa4b", 00:28:59.144 "is_configured": true, 00:28:59.144 "data_offset": 0, 00:28:59.144 "data_size": 65536 00:28:59.144 }, 00:28:59.144 { 00:28:59.144 "name": null, 00:28:59.144 "uuid": "989e0202-545d-4fc7-848f-1f98f8fdd9b6", 00:28:59.144 "is_configured": false, 00:28:59.144 "data_offset": 0, 00:28:59.144 "data_size": 65536 00:28:59.144 }, 00:28:59.144 { 00:28:59.144 "name": null, 00:28:59.144 "uuid": "52a30748-1767-45e5-9109-cabee6bdc9a8", 00:28:59.144 "is_configured": false, 00:28:59.144 "data_offset": 0, 00:28:59.144 "data_size": 65536 00:28:59.144 }, 00:28:59.144 { 00:28:59.144 "name": "BaseBdev4", 00:28:59.144 "uuid": "648c9b30-6d19-47e6-a202-109991a1150a", 00:28:59.144 "is_configured": true, 00:28:59.144 "data_offset": 0, 00:28:59.144 "data_size": 65536 00:28:59.144 } 00:28:59.144 ] 00:28:59.144 }' 00:28:59.144 13:49:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:59.144 13:49:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:59.403 [2024-11-20 13:49:06.302632] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:59.403 "name": "Existed_Raid", 00:28:59.403 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:59.403 "strip_size_kb": 64, 00:28:59.403 "state": "configuring", 00:28:59.403 "raid_level": "raid0", 00:28:59.403 "superblock": false, 00:28:59.403 "num_base_bdevs": 4, 00:28:59.403 "num_base_bdevs_discovered": 3, 00:28:59.403 "num_base_bdevs_operational": 4, 00:28:59.403 "base_bdevs_list": [ 00:28:59.403 { 00:28:59.403 "name": "BaseBdev1", 00:28:59.403 "uuid": "24d68baa-7898-45c7-9b41-8957c641fa4b", 00:28:59.403 "is_configured": true, 00:28:59.403 "data_offset": 0, 00:28:59.403 "data_size": 65536 00:28:59.403 }, 00:28:59.403 { 00:28:59.403 "name": null, 00:28:59.403 "uuid": "989e0202-545d-4fc7-848f-1f98f8fdd9b6", 00:28:59.403 "is_configured": false, 00:28:59.403 "data_offset": 0, 00:28:59.403 "data_size": 65536 00:28:59.403 }, 00:28:59.403 { 00:28:59.403 "name": "BaseBdev3", 00:28:59.403 "uuid": "52a30748-1767-45e5-9109-cabee6bdc9a8", 00:28:59.403 "is_configured": true, 00:28:59.403 "data_offset": 0, 00:28:59.403 "data_size": 65536 00:28:59.403 }, 00:28:59.403 { 00:28:59.403 "name": "BaseBdev4", 00:28:59.403 "uuid": "648c9b30-6d19-47e6-a202-109991a1150a", 00:28:59.403 "is_configured": true, 00:28:59.403 "data_offset": 0, 00:28:59.403 "data_size": 65536 00:28:59.403 } 00:28:59.403 ] 00:28:59.403 }' 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:59.403 13:49:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:59.662 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:59.662 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:28:59.662 13:49:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:59.662 13:49:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:59.662 13:49:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:59.662 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:28:59.662 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:28:59.662 13:49:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:59.662 13:49:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:59.662 [2024-11-20 13:49:06.670755] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:28:59.920 13:49:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:59.920 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:28:59.920 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:59.921 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:59.921 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:28:59.921 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:59.921 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:28:59.921 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:59.921 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:59.921 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:59.921 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:59.921 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:59.921 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:59.921 13:49:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:59.921 13:49:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:59.921 13:49:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:59.921 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:59.921 "name": "Existed_Raid", 00:28:59.921 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:59.921 "strip_size_kb": 64, 00:28:59.921 "state": "configuring", 00:28:59.921 "raid_level": "raid0", 00:28:59.921 "superblock": false, 00:28:59.921 "num_base_bdevs": 4, 00:28:59.921 "num_base_bdevs_discovered": 2, 00:28:59.921 "num_base_bdevs_operational": 4, 00:28:59.921 "base_bdevs_list": [ 00:28:59.921 { 00:28:59.921 "name": null, 00:28:59.921 "uuid": "24d68baa-7898-45c7-9b41-8957c641fa4b", 00:28:59.921 "is_configured": false, 00:28:59.921 "data_offset": 0, 00:28:59.921 "data_size": 65536 00:28:59.921 }, 00:28:59.921 { 00:28:59.921 "name": null, 00:28:59.921 "uuid": "989e0202-545d-4fc7-848f-1f98f8fdd9b6", 00:28:59.921 "is_configured": false, 00:28:59.921 "data_offset": 0, 00:28:59.921 "data_size": 65536 00:28:59.921 }, 00:28:59.921 { 00:28:59.921 "name": "BaseBdev3", 00:28:59.921 "uuid": "52a30748-1767-45e5-9109-cabee6bdc9a8", 00:28:59.921 "is_configured": true, 00:28:59.921 "data_offset": 0, 00:28:59.921 "data_size": 65536 00:28:59.921 }, 00:28:59.921 { 00:28:59.921 "name": "BaseBdev4", 00:28:59.921 "uuid": "648c9b30-6d19-47e6-a202-109991a1150a", 00:28:59.921 "is_configured": true, 00:28:59.921 "data_offset": 0, 00:28:59.921 "data_size": 65536 00:28:59.921 } 00:28:59.921 ] 00:28:59.921 }' 00:28:59.921 13:49:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:59.921 13:49:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:00.179 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:00.179 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:00.180 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:00.180 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:29:00.180 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:00.180 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:29:00.180 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:29:00.180 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:00.180 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:00.180 [2024-11-20 13:49:07.080141] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:29:00.180 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:00.180 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:29:00.180 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:00.180 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:00.180 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:29:00.180 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:00.180 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:00.180 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:00.180 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:00.180 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:00.180 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:00.180 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:00.180 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:00.180 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:00.180 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:00.180 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:00.180 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:00.180 "name": "Existed_Raid", 00:29:00.180 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:00.180 "strip_size_kb": 64, 00:29:00.180 "state": "configuring", 00:29:00.180 "raid_level": "raid0", 00:29:00.180 "superblock": false, 00:29:00.180 "num_base_bdevs": 4, 00:29:00.180 "num_base_bdevs_discovered": 3, 00:29:00.180 "num_base_bdevs_operational": 4, 00:29:00.180 "base_bdevs_list": [ 00:29:00.180 { 00:29:00.180 "name": null, 00:29:00.180 "uuid": "24d68baa-7898-45c7-9b41-8957c641fa4b", 00:29:00.180 "is_configured": false, 00:29:00.180 "data_offset": 0, 00:29:00.180 "data_size": 65536 00:29:00.180 }, 00:29:00.180 { 00:29:00.180 "name": "BaseBdev2", 00:29:00.180 "uuid": "989e0202-545d-4fc7-848f-1f98f8fdd9b6", 00:29:00.180 "is_configured": true, 00:29:00.180 "data_offset": 0, 00:29:00.180 "data_size": 65536 00:29:00.180 }, 00:29:00.180 { 00:29:00.180 "name": "BaseBdev3", 00:29:00.180 "uuid": "52a30748-1767-45e5-9109-cabee6bdc9a8", 00:29:00.180 "is_configured": true, 00:29:00.180 "data_offset": 0, 00:29:00.180 "data_size": 65536 00:29:00.180 }, 00:29:00.180 { 00:29:00.180 "name": "BaseBdev4", 00:29:00.180 "uuid": "648c9b30-6d19-47e6-a202-109991a1150a", 00:29:00.180 "is_configured": true, 00:29:00.180 "data_offset": 0, 00:29:00.180 "data_size": 65536 00:29:00.180 } 00:29:00.180 ] 00:29:00.180 }' 00:29:00.180 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:00.180 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:00.437 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:00.437 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:29:00.437 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:00.437 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:00.437 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:00.437 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:29:00.437 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:00.437 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:29:00.437 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:00.437 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:00.437 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 24d68baa-7898-45c7-9b41-8957c641fa4b 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:00.696 [2024-11-20 13:49:07.524539] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:29:00.696 [2024-11-20 13:49:07.524693] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:29:00.696 [2024-11-20 13:49:07.524706] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:29:00.696 [2024-11-20 13:49:07.524960] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:29:00.696 [2024-11-20 13:49:07.525103] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:29:00.696 [2024-11-20 13:49:07.525113] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:29:00.696 [2024-11-20 13:49:07.525315] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:00.696 NewBaseBdev 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:00.696 [ 00:29:00.696 { 00:29:00.696 "name": "NewBaseBdev", 00:29:00.696 "aliases": [ 00:29:00.696 "24d68baa-7898-45c7-9b41-8957c641fa4b" 00:29:00.696 ], 00:29:00.696 "product_name": "Malloc disk", 00:29:00.696 "block_size": 512, 00:29:00.696 "num_blocks": 65536, 00:29:00.696 "uuid": "24d68baa-7898-45c7-9b41-8957c641fa4b", 00:29:00.696 "assigned_rate_limits": { 00:29:00.696 "rw_ios_per_sec": 0, 00:29:00.696 "rw_mbytes_per_sec": 0, 00:29:00.696 "r_mbytes_per_sec": 0, 00:29:00.696 "w_mbytes_per_sec": 0 00:29:00.696 }, 00:29:00.696 "claimed": true, 00:29:00.696 "claim_type": "exclusive_write", 00:29:00.696 "zoned": false, 00:29:00.696 "supported_io_types": { 00:29:00.696 "read": true, 00:29:00.696 "write": true, 00:29:00.696 "unmap": true, 00:29:00.696 "flush": true, 00:29:00.696 "reset": true, 00:29:00.696 "nvme_admin": false, 00:29:00.696 "nvme_io": false, 00:29:00.696 "nvme_io_md": false, 00:29:00.696 "write_zeroes": true, 00:29:00.696 "zcopy": true, 00:29:00.696 "get_zone_info": false, 00:29:00.696 "zone_management": false, 00:29:00.696 "zone_append": false, 00:29:00.696 "compare": false, 00:29:00.696 "compare_and_write": false, 00:29:00.696 "abort": true, 00:29:00.696 "seek_hole": false, 00:29:00.696 "seek_data": false, 00:29:00.696 "copy": true, 00:29:00.696 "nvme_iov_md": false 00:29:00.696 }, 00:29:00.696 "memory_domains": [ 00:29:00.696 { 00:29:00.696 "dma_device_id": "system", 00:29:00.696 "dma_device_type": 1 00:29:00.696 }, 00:29:00.696 { 00:29:00.696 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:00.696 "dma_device_type": 2 00:29:00.696 } 00:29:00.696 ], 00:29:00.696 "driver_specific": {} 00:29:00.696 } 00:29:00.696 ] 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:00.696 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:00.696 "name": "Existed_Raid", 00:29:00.696 "uuid": "d41e8d4d-52c2-404f-8bd7-7be7d8caf96b", 00:29:00.696 "strip_size_kb": 64, 00:29:00.696 "state": "online", 00:29:00.696 "raid_level": "raid0", 00:29:00.696 "superblock": false, 00:29:00.696 "num_base_bdevs": 4, 00:29:00.696 "num_base_bdevs_discovered": 4, 00:29:00.696 "num_base_bdevs_operational": 4, 00:29:00.696 "base_bdevs_list": [ 00:29:00.696 { 00:29:00.696 "name": "NewBaseBdev", 00:29:00.696 "uuid": "24d68baa-7898-45c7-9b41-8957c641fa4b", 00:29:00.696 "is_configured": true, 00:29:00.696 "data_offset": 0, 00:29:00.696 "data_size": 65536 00:29:00.696 }, 00:29:00.696 { 00:29:00.696 "name": "BaseBdev2", 00:29:00.697 "uuid": "989e0202-545d-4fc7-848f-1f98f8fdd9b6", 00:29:00.697 "is_configured": true, 00:29:00.697 "data_offset": 0, 00:29:00.697 "data_size": 65536 00:29:00.697 }, 00:29:00.697 { 00:29:00.697 "name": "BaseBdev3", 00:29:00.697 "uuid": "52a30748-1767-45e5-9109-cabee6bdc9a8", 00:29:00.697 "is_configured": true, 00:29:00.697 "data_offset": 0, 00:29:00.697 "data_size": 65536 00:29:00.697 }, 00:29:00.697 { 00:29:00.697 "name": "BaseBdev4", 00:29:00.697 "uuid": "648c9b30-6d19-47e6-a202-109991a1150a", 00:29:00.697 "is_configured": true, 00:29:00.697 "data_offset": 0, 00:29:00.697 "data_size": 65536 00:29:00.697 } 00:29:00.697 ] 00:29:00.697 }' 00:29:00.697 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:00.697 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:00.955 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:29:00.955 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:29:00.955 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:29:00.955 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:29:00.955 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:29:00.955 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:29:00.955 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:29:00.955 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:29:00.955 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:00.955 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:00.955 [2024-11-20 13:49:07.869025] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:00.955 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:00.955 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:00.955 "name": "Existed_Raid", 00:29:00.955 "aliases": [ 00:29:00.955 "d41e8d4d-52c2-404f-8bd7-7be7d8caf96b" 00:29:00.955 ], 00:29:00.955 "product_name": "Raid Volume", 00:29:00.955 "block_size": 512, 00:29:00.955 "num_blocks": 262144, 00:29:00.955 "uuid": "d41e8d4d-52c2-404f-8bd7-7be7d8caf96b", 00:29:00.955 "assigned_rate_limits": { 00:29:00.955 "rw_ios_per_sec": 0, 00:29:00.955 "rw_mbytes_per_sec": 0, 00:29:00.955 "r_mbytes_per_sec": 0, 00:29:00.955 "w_mbytes_per_sec": 0 00:29:00.955 }, 00:29:00.955 "claimed": false, 00:29:00.955 "zoned": false, 00:29:00.955 "supported_io_types": { 00:29:00.955 "read": true, 00:29:00.955 "write": true, 00:29:00.955 "unmap": true, 00:29:00.955 "flush": true, 00:29:00.955 "reset": true, 00:29:00.955 "nvme_admin": false, 00:29:00.955 "nvme_io": false, 00:29:00.955 "nvme_io_md": false, 00:29:00.955 "write_zeroes": true, 00:29:00.955 "zcopy": false, 00:29:00.955 "get_zone_info": false, 00:29:00.955 "zone_management": false, 00:29:00.955 "zone_append": false, 00:29:00.955 "compare": false, 00:29:00.955 "compare_and_write": false, 00:29:00.955 "abort": false, 00:29:00.955 "seek_hole": false, 00:29:00.955 "seek_data": false, 00:29:00.955 "copy": false, 00:29:00.955 "nvme_iov_md": false 00:29:00.955 }, 00:29:00.955 "memory_domains": [ 00:29:00.955 { 00:29:00.955 "dma_device_id": "system", 00:29:00.955 "dma_device_type": 1 00:29:00.955 }, 00:29:00.955 { 00:29:00.955 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:00.955 "dma_device_type": 2 00:29:00.955 }, 00:29:00.955 { 00:29:00.955 "dma_device_id": "system", 00:29:00.955 "dma_device_type": 1 00:29:00.955 }, 00:29:00.955 { 00:29:00.955 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:00.955 "dma_device_type": 2 00:29:00.955 }, 00:29:00.955 { 00:29:00.955 "dma_device_id": "system", 00:29:00.955 "dma_device_type": 1 00:29:00.955 }, 00:29:00.955 { 00:29:00.955 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:00.955 "dma_device_type": 2 00:29:00.955 }, 00:29:00.955 { 00:29:00.955 "dma_device_id": "system", 00:29:00.955 "dma_device_type": 1 00:29:00.955 }, 00:29:00.955 { 00:29:00.955 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:00.955 "dma_device_type": 2 00:29:00.955 } 00:29:00.955 ], 00:29:00.955 "driver_specific": { 00:29:00.955 "raid": { 00:29:00.955 "uuid": "d41e8d4d-52c2-404f-8bd7-7be7d8caf96b", 00:29:00.955 "strip_size_kb": 64, 00:29:00.955 "state": "online", 00:29:00.955 "raid_level": "raid0", 00:29:00.955 "superblock": false, 00:29:00.955 "num_base_bdevs": 4, 00:29:00.955 "num_base_bdevs_discovered": 4, 00:29:00.955 "num_base_bdevs_operational": 4, 00:29:00.955 "base_bdevs_list": [ 00:29:00.955 { 00:29:00.955 "name": "NewBaseBdev", 00:29:00.955 "uuid": "24d68baa-7898-45c7-9b41-8957c641fa4b", 00:29:00.955 "is_configured": true, 00:29:00.955 "data_offset": 0, 00:29:00.955 "data_size": 65536 00:29:00.955 }, 00:29:00.955 { 00:29:00.955 "name": "BaseBdev2", 00:29:00.955 "uuid": "989e0202-545d-4fc7-848f-1f98f8fdd9b6", 00:29:00.955 "is_configured": true, 00:29:00.955 "data_offset": 0, 00:29:00.955 "data_size": 65536 00:29:00.955 }, 00:29:00.955 { 00:29:00.955 "name": "BaseBdev3", 00:29:00.955 "uuid": "52a30748-1767-45e5-9109-cabee6bdc9a8", 00:29:00.955 "is_configured": true, 00:29:00.955 "data_offset": 0, 00:29:00.955 "data_size": 65536 00:29:00.955 }, 00:29:00.955 { 00:29:00.955 "name": "BaseBdev4", 00:29:00.955 "uuid": "648c9b30-6d19-47e6-a202-109991a1150a", 00:29:00.955 "is_configured": true, 00:29:00.955 "data_offset": 0, 00:29:00.955 "data_size": 65536 00:29:00.955 } 00:29:00.955 ] 00:29:00.955 } 00:29:00.955 } 00:29:00.955 }' 00:29:00.955 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:29:00.955 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:29:00.955 BaseBdev2 00:29:00.955 BaseBdev3 00:29:00.955 BaseBdev4' 00:29:00.955 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:00.955 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:29:00.955 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:00.955 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:29:00.955 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:00.955 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:00.955 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:00.956 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:00.956 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:00.956 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:00.956 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:00.956 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:29:00.956 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:00.956 13:49:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:00.956 13:49:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:01.214 [2024-11-20 13:49:08.100716] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:29:01.214 [2024-11-20 13:49:08.100750] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:29:01.214 [2024-11-20 13:49:08.100828] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:01.214 [2024-11-20 13:49:08.100894] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:29:01.214 [2024-11-20 13:49:08.100903] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 67609 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 67609 ']' 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 67609 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 67609 00:29:01.214 killing process with pid 67609 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 67609' 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 67609 00:29:01.214 [2024-11-20 13:49:08.126916] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:29:01.214 13:49:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 67609 00:29:01.472 [2024-11-20 13:49:08.331675] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:29:02.039 ************************************ 00:29:02.039 END TEST raid_state_function_test 00:29:02.039 ************************************ 00:29:02.039 13:49:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:29:02.039 00:29:02.039 real 0m8.569s 00:29:02.039 user 0m13.783s 00:29:02.039 sys 0m1.438s 00:29:02.039 13:49:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:29:02.039 13:49:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:02.039 13:49:09 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 4 true 00:29:02.039 13:49:09 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:29:02.039 13:49:09 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:29:02.039 13:49:09 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:29:02.039 ************************************ 00:29:02.039 START TEST raid_state_function_test_sb 00:29:02.039 ************************************ 00:29:02.039 13:49:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 4 true 00:29:02.039 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:29:02.039 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:29:02.039 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:29:02.039 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:29:02.039 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:29:02.039 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:02.039 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:29:02.039 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:29:02.039 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:02.039 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:29:02.039 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:29:02.039 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:02.039 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:29:02.039 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:29:02.040 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:02.040 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:29:02.040 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:29:02.040 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:02.040 Process raid pid: 68253 00:29:02.040 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:02.040 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:29:02.040 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:29:02.040 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:29:02.040 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:29:02.040 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:29:02.040 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:29:02.040 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:29:02.040 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:29:02.040 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:29:02.040 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:29:02.040 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:29:02.040 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=68253 00:29:02.040 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 68253' 00:29:02.040 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 68253 00:29:02.040 13:49:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 68253 ']' 00:29:02.040 13:49:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:02.040 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:29:02.040 13:49:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:29:02.040 13:49:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:02.040 13:49:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:29:02.040 13:49:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:02.298 [2024-11-20 13:49:09.103463] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:29:02.298 [2024-11-20 13:49:09.103605] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:29:02.298 [2024-11-20 13:49:09.260794] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:02.556 [2024-11-20 13:49:09.382157] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:29:02.556 [2024-11-20 13:49:09.532725] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:02.556 [2024-11-20 13:49:09.532939] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:03.121 13:49:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:29:03.121 13:49:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:29:03.121 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:29:03.121 13:49:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:03.121 13:49:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:03.121 [2024-11-20 13:49:09.957006] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:29:03.121 [2024-11-20 13:49:09.957080] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:29:03.121 [2024-11-20 13:49:09.957091] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:29:03.121 [2024-11-20 13:49:09.957101] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:29:03.121 [2024-11-20 13:49:09.957108] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:29:03.121 [2024-11-20 13:49:09.957116] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:29:03.121 [2024-11-20 13:49:09.957123] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:29:03.121 [2024-11-20 13:49:09.957132] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:29:03.121 13:49:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:03.121 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:29:03.121 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:03.121 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:03.121 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:29:03.121 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:03.121 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:03.121 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:03.121 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:03.121 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:03.121 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:03.121 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:03.121 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:03.121 13:49:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:03.121 13:49:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:03.121 13:49:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:03.121 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:03.121 "name": "Existed_Raid", 00:29:03.121 "uuid": "1b8896a4-4e20-4da9-8540-70be7421d268", 00:29:03.121 "strip_size_kb": 64, 00:29:03.121 "state": "configuring", 00:29:03.121 "raid_level": "raid0", 00:29:03.121 "superblock": true, 00:29:03.121 "num_base_bdevs": 4, 00:29:03.121 "num_base_bdevs_discovered": 0, 00:29:03.121 "num_base_bdevs_operational": 4, 00:29:03.121 "base_bdevs_list": [ 00:29:03.121 { 00:29:03.121 "name": "BaseBdev1", 00:29:03.121 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:03.121 "is_configured": false, 00:29:03.121 "data_offset": 0, 00:29:03.121 "data_size": 0 00:29:03.121 }, 00:29:03.121 { 00:29:03.121 "name": "BaseBdev2", 00:29:03.121 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:03.121 "is_configured": false, 00:29:03.121 "data_offset": 0, 00:29:03.121 "data_size": 0 00:29:03.121 }, 00:29:03.121 { 00:29:03.121 "name": "BaseBdev3", 00:29:03.121 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:03.121 "is_configured": false, 00:29:03.121 "data_offset": 0, 00:29:03.121 "data_size": 0 00:29:03.121 }, 00:29:03.121 { 00:29:03.121 "name": "BaseBdev4", 00:29:03.121 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:03.121 "is_configured": false, 00:29:03.121 "data_offset": 0, 00:29:03.121 "data_size": 0 00:29:03.121 } 00:29:03.121 ] 00:29:03.121 }' 00:29:03.121 13:49:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:03.121 13:49:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:03.380 [2024-11-20 13:49:10.317020] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:29:03.380 [2024-11-20 13:49:10.317064] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:03.380 [2024-11-20 13:49:10.325013] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:29:03.380 [2024-11-20 13:49:10.325140] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:29:03.380 [2024-11-20 13:49:10.325203] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:29:03.380 [2024-11-20 13:49:10.325232] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:29:03.380 [2024-11-20 13:49:10.325251] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:29:03.380 [2024-11-20 13:49:10.325272] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:29:03.380 [2024-11-20 13:49:10.325330] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:29:03.380 [2024-11-20 13:49:10.325356] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:03.380 [2024-11-20 13:49:10.359879] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:29:03.380 BaseBdev1 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:03.380 [ 00:29:03.380 { 00:29:03.380 "name": "BaseBdev1", 00:29:03.380 "aliases": [ 00:29:03.380 "2c3a63e0-ccad-48fa-9aa8-c1e599609c3d" 00:29:03.380 ], 00:29:03.380 "product_name": "Malloc disk", 00:29:03.380 "block_size": 512, 00:29:03.380 "num_blocks": 65536, 00:29:03.380 "uuid": "2c3a63e0-ccad-48fa-9aa8-c1e599609c3d", 00:29:03.380 "assigned_rate_limits": { 00:29:03.380 "rw_ios_per_sec": 0, 00:29:03.380 "rw_mbytes_per_sec": 0, 00:29:03.380 "r_mbytes_per_sec": 0, 00:29:03.380 "w_mbytes_per_sec": 0 00:29:03.380 }, 00:29:03.380 "claimed": true, 00:29:03.380 "claim_type": "exclusive_write", 00:29:03.380 "zoned": false, 00:29:03.380 "supported_io_types": { 00:29:03.380 "read": true, 00:29:03.380 "write": true, 00:29:03.380 "unmap": true, 00:29:03.380 "flush": true, 00:29:03.380 "reset": true, 00:29:03.380 "nvme_admin": false, 00:29:03.380 "nvme_io": false, 00:29:03.380 "nvme_io_md": false, 00:29:03.380 "write_zeroes": true, 00:29:03.380 "zcopy": true, 00:29:03.380 "get_zone_info": false, 00:29:03.380 "zone_management": false, 00:29:03.380 "zone_append": false, 00:29:03.380 "compare": false, 00:29:03.380 "compare_and_write": false, 00:29:03.380 "abort": true, 00:29:03.380 "seek_hole": false, 00:29:03.380 "seek_data": false, 00:29:03.380 "copy": true, 00:29:03.380 "nvme_iov_md": false 00:29:03.380 }, 00:29:03.380 "memory_domains": [ 00:29:03.380 { 00:29:03.380 "dma_device_id": "system", 00:29:03.380 "dma_device_type": 1 00:29:03.380 }, 00:29:03.380 { 00:29:03.380 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:03.380 "dma_device_type": 2 00:29:03.380 } 00:29:03.380 ], 00:29:03.380 "driver_specific": {} 00:29:03.380 } 00:29:03.380 ] 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:03.380 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:03.380 "name": "Existed_Raid", 00:29:03.380 "uuid": "6abb6b6e-d879-4ef9-bf0f-6bf5db21445e", 00:29:03.380 "strip_size_kb": 64, 00:29:03.380 "state": "configuring", 00:29:03.380 "raid_level": "raid0", 00:29:03.381 "superblock": true, 00:29:03.381 "num_base_bdevs": 4, 00:29:03.381 "num_base_bdevs_discovered": 1, 00:29:03.381 "num_base_bdevs_operational": 4, 00:29:03.381 "base_bdevs_list": [ 00:29:03.381 { 00:29:03.381 "name": "BaseBdev1", 00:29:03.381 "uuid": "2c3a63e0-ccad-48fa-9aa8-c1e599609c3d", 00:29:03.381 "is_configured": true, 00:29:03.381 "data_offset": 2048, 00:29:03.381 "data_size": 63488 00:29:03.381 }, 00:29:03.381 { 00:29:03.381 "name": "BaseBdev2", 00:29:03.381 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:03.381 "is_configured": false, 00:29:03.381 "data_offset": 0, 00:29:03.381 "data_size": 0 00:29:03.381 }, 00:29:03.381 { 00:29:03.381 "name": "BaseBdev3", 00:29:03.381 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:03.381 "is_configured": false, 00:29:03.381 "data_offset": 0, 00:29:03.381 "data_size": 0 00:29:03.381 }, 00:29:03.381 { 00:29:03.381 "name": "BaseBdev4", 00:29:03.381 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:03.381 "is_configured": false, 00:29:03.381 "data_offset": 0, 00:29:03.381 "data_size": 0 00:29:03.381 } 00:29:03.381 ] 00:29:03.381 }' 00:29:03.381 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:03.381 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:04.018 [2024-11-20 13:49:10.708033] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:29:04.018 [2024-11-20 13:49:10.708091] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:04.018 [2024-11-20 13:49:10.716099] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:29:04.018 [2024-11-20 13:49:10.718139] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:29:04.018 [2024-11-20 13:49:10.718177] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:29:04.018 [2024-11-20 13:49:10.718187] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:29:04.018 [2024-11-20 13:49:10.718198] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:29:04.018 [2024-11-20 13:49:10.718204] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:29:04.018 [2024-11-20 13:49:10.718215] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:04.018 "name": "Existed_Raid", 00:29:04.018 "uuid": "c0e47f68-95ea-462a-be62-b9ed13b395f0", 00:29:04.018 "strip_size_kb": 64, 00:29:04.018 "state": "configuring", 00:29:04.018 "raid_level": "raid0", 00:29:04.018 "superblock": true, 00:29:04.018 "num_base_bdevs": 4, 00:29:04.018 "num_base_bdevs_discovered": 1, 00:29:04.018 "num_base_bdevs_operational": 4, 00:29:04.018 "base_bdevs_list": [ 00:29:04.018 { 00:29:04.018 "name": "BaseBdev1", 00:29:04.018 "uuid": "2c3a63e0-ccad-48fa-9aa8-c1e599609c3d", 00:29:04.018 "is_configured": true, 00:29:04.018 "data_offset": 2048, 00:29:04.018 "data_size": 63488 00:29:04.018 }, 00:29:04.018 { 00:29:04.018 "name": "BaseBdev2", 00:29:04.018 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:04.018 "is_configured": false, 00:29:04.018 "data_offset": 0, 00:29:04.018 "data_size": 0 00:29:04.018 }, 00:29:04.018 { 00:29:04.018 "name": "BaseBdev3", 00:29:04.018 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:04.018 "is_configured": false, 00:29:04.018 "data_offset": 0, 00:29:04.018 "data_size": 0 00:29:04.018 }, 00:29:04.018 { 00:29:04.018 "name": "BaseBdev4", 00:29:04.018 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:04.018 "is_configured": false, 00:29:04.018 "data_offset": 0, 00:29:04.018 "data_size": 0 00:29:04.018 } 00:29:04.018 ] 00:29:04.018 }' 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:04.018 13:49:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:04.018 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:29:04.018 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:04.018 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:04.018 [2024-11-20 13:49:11.065038] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:29:04.018 BaseBdev2 00:29:04.018 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:04.018 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:29:04.018 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:29:04.018 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:04.018 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:29:04.018 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:04.018 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:04.018 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:04.018 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:04.018 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:04.278 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:04.278 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:29:04.278 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:04.278 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:04.278 [ 00:29:04.278 { 00:29:04.278 "name": "BaseBdev2", 00:29:04.278 "aliases": [ 00:29:04.278 "448e24d4-4b20-4e40-b2b6-adb97d216e65" 00:29:04.278 ], 00:29:04.278 "product_name": "Malloc disk", 00:29:04.278 "block_size": 512, 00:29:04.278 "num_blocks": 65536, 00:29:04.278 "uuid": "448e24d4-4b20-4e40-b2b6-adb97d216e65", 00:29:04.278 "assigned_rate_limits": { 00:29:04.278 "rw_ios_per_sec": 0, 00:29:04.278 "rw_mbytes_per_sec": 0, 00:29:04.278 "r_mbytes_per_sec": 0, 00:29:04.278 "w_mbytes_per_sec": 0 00:29:04.278 }, 00:29:04.278 "claimed": true, 00:29:04.278 "claim_type": "exclusive_write", 00:29:04.278 "zoned": false, 00:29:04.278 "supported_io_types": { 00:29:04.278 "read": true, 00:29:04.278 "write": true, 00:29:04.278 "unmap": true, 00:29:04.278 "flush": true, 00:29:04.278 "reset": true, 00:29:04.278 "nvme_admin": false, 00:29:04.278 "nvme_io": false, 00:29:04.278 "nvme_io_md": false, 00:29:04.278 "write_zeroes": true, 00:29:04.278 "zcopy": true, 00:29:04.278 "get_zone_info": false, 00:29:04.278 "zone_management": false, 00:29:04.278 "zone_append": false, 00:29:04.278 "compare": false, 00:29:04.278 "compare_and_write": false, 00:29:04.278 "abort": true, 00:29:04.278 "seek_hole": false, 00:29:04.278 "seek_data": false, 00:29:04.278 "copy": true, 00:29:04.278 "nvme_iov_md": false 00:29:04.278 }, 00:29:04.278 "memory_domains": [ 00:29:04.278 { 00:29:04.278 "dma_device_id": "system", 00:29:04.278 "dma_device_type": 1 00:29:04.278 }, 00:29:04.278 { 00:29:04.278 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:04.278 "dma_device_type": 2 00:29:04.278 } 00:29:04.278 ], 00:29:04.278 "driver_specific": {} 00:29:04.278 } 00:29:04.278 ] 00:29:04.278 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:04.278 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:29:04.278 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:29:04.278 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:29:04.278 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:29:04.278 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:04.278 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:04.278 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:29:04.278 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:04.278 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:04.278 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:04.278 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:04.278 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:04.278 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:04.278 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:04.278 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:04.278 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:04.278 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:04.278 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:04.278 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:04.278 "name": "Existed_Raid", 00:29:04.278 "uuid": "c0e47f68-95ea-462a-be62-b9ed13b395f0", 00:29:04.278 "strip_size_kb": 64, 00:29:04.278 "state": "configuring", 00:29:04.278 "raid_level": "raid0", 00:29:04.278 "superblock": true, 00:29:04.278 "num_base_bdevs": 4, 00:29:04.278 "num_base_bdevs_discovered": 2, 00:29:04.278 "num_base_bdevs_operational": 4, 00:29:04.278 "base_bdevs_list": [ 00:29:04.278 { 00:29:04.278 "name": "BaseBdev1", 00:29:04.278 "uuid": "2c3a63e0-ccad-48fa-9aa8-c1e599609c3d", 00:29:04.278 "is_configured": true, 00:29:04.278 "data_offset": 2048, 00:29:04.278 "data_size": 63488 00:29:04.278 }, 00:29:04.278 { 00:29:04.278 "name": "BaseBdev2", 00:29:04.278 "uuid": "448e24d4-4b20-4e40-b2b6-adb97d216e65", 00:29:04.278 "is_configured": true, 00:29:04.278 "data_offset": 2048, 00:29:04.278 "data_size": 63488 00:29:04.278 }, 00:29:04.278 { 00:29:04.278 "name": "BaseBdev3", 00:29:04.278 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:04.278 "is_configured": false, 00:29:04.278 "data_offset": 0, 00:29:04.278 "data_size": 0 00:29:04.278 }, 00:29:04.278 { 00:29:04.278 "name": "BaseBdev4", 00:29:04.278 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:04.278 "is_configured": false, 00:29:04.278 "data_offset": 0, 00:29:04.278 "data_size": 0 00:29:04.278 } 00:29:04.278 ] 00:29:04.278 }' 00:29:04.278 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:04.278 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:04.538 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:29:04.538 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:04.538 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:04.538 [2024-11-20 13:49:11.453321] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:29:04.538 BaseBdev3 00:29:04.538 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:04.538 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:29:04.538 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:29:04.538 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:04.538 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:29:04.538 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:04.538 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:04.538 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:04.538 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:04.538 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:04.538 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:04.538 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:29:04.538 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:04.538 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:04.538 [ 00:29:04.538 { 00:29:04.538 "name": "BaseBdev3", 00:29:04.538 "aliases": [ 00:29:04.538 "232214f1-0d72-47b0-9514-29cff39e288c" 00:29:04.538 ], 00:29:04.538 "product_name": "Malloc disk", 00:29:04.538 "block_size": 512, 00:29:04.538 "num_blocks": 65536, 00:29:04.538 "uuid": "232214f1-0d72-47b0-9514-29cff39e288c", 00:29:04.538 "assigned_rate_limits": { 00:29:04.539 "rw_ios_per_sec": 0, 00:29:04.539 "rw_mbytes_per_sec": 0, 00:29:04.539 "r_mbytes_per_sec": 0, 00:29:04.539 "w_mbytes_per_sec": 0 00:29:04.539 }, 00:29:04.539 "claimed": true, 00:29:04.539 "claim_type": "exclusive_write", 00:29:04.539 "zoned": false, 00:29:04.539 "supported_io_types": { 00:29:04.539 "read": true, 00:29:04.539 "write": true, 00:29:04.539 "unmap": true, 00:29:04.539 "flush": true, 00:29:04.539 "reset": true, 00:29:04.539 "nvme_admin": false, 00:29:04.539 "nvme_io": false, 00:29:04.539 "nvme_io_md": false, 00:29:04.539 "write_zeroes": true, 00:29:04.539 "zcopy": true, 00:29:04.539 "get_zone_info": false, 00:29:04.539 "zone_management": false, 00:29:04.539 "zone_append": false, 00:29:04.539 "compare": false, 00:29:04.539 "compare_and_write": false, 00:29:04.539 "abort": true, 00:29:04.539 "seek_hole": false, 00:29:04.539 "seek_data": false, 00:29:04.539 "copy": true, 00:29:04.539 "nvme_iov_md": false 00:29:04.539 }, 00:29:04.539 "memory_domains": [ 00:29:04.539 { 00:29:04.539 "dma_device_id": "system", 00:29:04.539 "dma_device_type": 1 00:29:04.539 }, 00:29:04.539 { 00:29:04.539 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:04.539 "dma_device_type": 2 00:29:04.539 } 00:29:04.539 ], 00:29:04.539 "driver_specific": {} 00:29:04.539 } 00:29:04.539 ] 00:29:04.539 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:04.539 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:29:04.539 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:29:04.539 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:29:04.539 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:29:04.539 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:04.539 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:04.539 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:29:04.539 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:04.539 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:04.539 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:04.539 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:04.539 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:04.539 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:04.539 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:04.539 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:04.539 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:04.539 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:04.539 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:04.539 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:04.539 "name": "Existed_Raid", 00:29:04.539 "uuid": "c0e47f68-95ea-462a-be62-b9ed13b395f0", 00:29:04.539 "strip_size_kb": 64, 00:29:04.539 "state": "configuring", 00:29:04.539 "raid_level": "raid0", 00:29:04.539 "superblock": true, 00:29:04.539 "num_base_bdevs": 4, 00:29:04.539 "num_base_bdevs_discovered": 3, 00:29:04.539 "num_base_bdevs_operational": 4, 00:29:04.539 "base_bdevs_list": [ 00:29:04.539 { 00:29:04.539 "name": "BaseBdev1", 00:29:04.539 "uuid": "2c3a63e0-ccad-48fa-9aa8-c1e599609c3d", 00:29:04.539 "is_configured": true, 00:29:04.539 "data_offset": 2048, 00:29:04.539 "data_size": 63488 00:29:04.539 }, 00:29:04.539 { 00:29:04.539 "name": "BaseBdev2", 00:29:04.539 "uuid": "448e24d4-4b20-4e40-b2b6-adb97d216e65", 00:29:04.539 "is_configured": true, 00:29:04.539 "data_offset": 2048, 00:29:04.539 "data_size": 63488 00:29:04.539 }, 00:29:04.539 { 00:29:04.539 "name": "BaseBdev3", 00:29:04.539 "uuid": "232214f1-0d72-47b0-9514-29cff39e288c", 00:29:04.539 "is_configured": true, 00:29:04.539 "data_offset": 2048, 00:29:04.539 "data_size": 63488 00:29:04.539 }, 00:29:04.539 { 00:29:04.539 "name": "BaseBdev4", 00:29:04.539 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:04.539 "is_configured": false, 00:29:04.539 "data_offset": 0, 00:29:04.539 "data_size": 0 00:29:04.539 } 00:29:04.539 ] 00:29:04.539 }' 00:29:04.539 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:04.539 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:04.798 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:29:04.798 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:04.798 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:04.798 [2024-11-20 13:49:11.826399] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:29:04.798 [2024-11-20 13:49:11.826667] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:29:04.798 [2024-11-20 13:49:11.826682] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:29:04.798 [2024-11-20 13:49:11.826965] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:29:04.798 BaseBdev4 00:29:04.798 [2024-11-20 13:49:11.827131] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:29:04.798 [2024-11-20 13:49:11.827143] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:29:04.798 [2024-11-20 13:49:11.827300] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:04.798 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:04.798 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:29:04.798 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:29:04.798 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:04.798 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:29:04.798 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:04.798 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:04.798 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:04.798 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:04.798 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:04.798 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:04.798 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:29:04.798 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:04.798 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:04.798 [ 00:29:04.798 { 00:29:04.798 "name": "BaseBdev4", 00:29:04.798 "aliases": [ 00:29:04.798 "3d08640e-ebf3-4102-8c46-92d933ee30af" 00:29:04.798 ], 00:29:04.798 "product_name": "Malloc disk", 00:29:04.798 "block_size": 512, 00:29:04.798 "num_blocks": 65536, 00:29:04.798 "uuid": "3d08640e-ebf3-4102-8c46-92d933ee30af", 00:29:04.798 "assigned_rate_limits": { 00:29:04.798 "rw_ios_per_sec": 0, 00:29:04.798 "rw_mbytes_per_sec": 0, 00:29:04.798 "r_mbytes_per_sec": 0, 00:29:04.798 "w_mbytes_per_sec": 0 00:29:04.798 }, 00:29:04.798 "claimed": true, 00:29:04.798 "claim_type": "exclusive_write", 00:29:04.798 "zoned": false, 00:29:04.798 "supported_io_types": { 00:29:04.798 "read": true, 00:29:04.798 "write": true, 00:29:04.798 "unmap": true, 00:29:04.798 "flush": true, 00:29:04.798 "reset": true, 00:29:04.798 "nvme_admin": false, 00:29:04.798 "nvme_io": false, 00:29:04.798 "nvme_io_md": false, 00:29:04.798 "write_zeroes": true, 00:29:04.798 "zcopy": true, 00:29:04.798 "get_zone_info": false, 00:29:04.798 "zone_management": false, 00:29:04.798 "zone_append": false, 00:29:04.798 "compare": false, 00:29:04.798 "compare_and_write": false, 00:29:04.798 "abort": true, 00:29:04.798 "seek_hole": false, 00:29:04.798 "seek_data": false, 00:29:04.798 "copy": true, 00:29:04.798 "nvme_iov_md": false 00:29:04.798 }, 00:29:04.798 "memory_domains": [ 00:29:04.798 { 00:29:04.798 "dma_device_id": "system", 00:29:04.798 "dma_device_type": 1 00:29:04.798 }, 00:29:04.798 { 00:29:04.798 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:04.798 "dma_device_type": 2 00:29:04.798 } 00:29:04.798 ], 00:29:04.798 "driver_specific": {} 00:29:04.798 } 00:29:04.798 ] 00:29:04.798 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:04.798 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:29:04.798 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:29:04.798 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:29:04.798 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:29:04.798 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:04.798 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:04.798 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:29:04.798 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:05.057 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:05.057 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:05.057 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:05.057 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:05.057 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:05.057 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:05.057 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:05.057 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:05.057 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:05.057 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:05.057 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:05.057 "name": "Existed_Raid", 00:29:05.057 "uuid": "c0e47f68-95ea-462a-be62-b9ed13b395f0", 00:29:05.057 "strip_size_kb": 64, 00:29:05.057 "state": "online", 00:29:05.057 "raid_level": "raid0", 00:29:05.057 "superblock": true, 00:29:05.057 "num_base_bdevs": 4, 00:29:05.057 "num_base_bdevs_discovered": 4, 00:29:05.057 "num_base_bdevs_operational": 4, 00:29:05.057 "base_bdevs_list": [ 00:29:05.057 { 00:29:05.057 "name": "BaseBdev1", 00:29:05.057 "uuid": "2c3a63e0-ccad-48fa-9aa8-c1e599609c3d", 00:29:05.057 "is_configured": true, 00:29:05.057 "data_offset": 2048, 00:29:05.057 "data_size": 63488 00:29:05.057 }, 00:29:05.057 { 00:29:05.057 "name": "BaseBdev2", 00:29:05.057 "uuid": "448e24d4-4b20-4e40-b2b6-adb97d216e65", 00:29:05.057 "is_configured": true, 00:29:05.057 "data_offset": 2048, 00:29:05.057 "data_size": 63488 00:29:05.057 }, 00:29:05.057 { 00:29:05.057 "name": "BaseBdev3", 00:29:05.057 "uuid": "232214f1-0d72-47b0-9514-29cff39e288c", 00:29:05.057 "is_configured": true, 00:29:05.057 "data_offset": 2048, 00:29:05.057 "data_size": 63488 00:29:05.057 }, 00:29:05.057 { 00:29:05.057 "name": "BaseBdev4", 00:29:05.057 "uuid": "3d08640e-ebf3-4102-8c46-92d933ee30af", 00:29:05.057 "is_configured": true, 00:29:05.057 "data_offset": 2048, 00:29:05.057 "data_size": 63488 00:29:05.057 } 00:29:05.057 ] 00:29:05.057 }' 00:29:05.057 13:49:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:05.057 13:49:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:05.317 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:29:05.317 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:29:05.317 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:29:05.317 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:29:05.317 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:29:05.317 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:29:05.317 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:29:05.317 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:29:05.317 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:05.317 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:05.317 [2024-11-20 13:49:12.178933] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:05.317 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:05.317 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:05.317 "name": "Existed_Raid", 00:29:05.317 "aliases": [ 00:29:05.317 "c0e47f68-95ea-462a-be62-b9ed13b395f0" 00:29:05.317 ], 00:29:05.317 "product_name": "Raid Volume", 00:29:05.317 "block_size": 512, 00:29:05.317 "num_blocks": 253952, 00:29:05.317 "uuid": "c0e47f68-95ea-462a-be62-b9ed13b395f0", 00:29:05.318 "assigned_rate_limits": { 00:29:05.318 "rw_ios_per_sec": 0, 00:29:05.318 "rw_mbytes_per_sec": 0, 00:29:05.318 "r_mbytes_per_sec": 0, 00:29:05.318 "w_mbytes_per_sec": 0 00:29:05.318 }, 00:29:05.318 "claimed": false, 00:29:05.318 "zoned": false, 00:29:05.318 "supported_io_types": { 00:29:05.318 "read": true, 00:29:05.318 "write": true, 00:29:05.318 "unmap": true, 00:29:05.318 "flush": true, 00:29:05.318 "reset": true, 00:29:05.318 "nvme_admin": false, 00:29:05.318 "nvme_io": false, 00:29:05.318 "nvme_io_md": false, 00:29:05.318 "write_zeroes": true, 00:29:05.318 "zcopy": false, 00:29:05.318 "get_zone_info": false, 00:29:05.318 "zone_management": false, 00:29:05.318 "zone_append": false, 00:29:05.318 "compare": false, 00:29:05.318 "compare_and_write": false, 00:29:05.318 "abort": false, 00:29:05.318 "seek_hole": false, 00:29:05.318 "seek_data": false, 00:29:05.318 "copy": false, 00:29:05.318 "nvme_iov_md": false 00:29:05.318 }, 00:29:05.318 "memory_domains": [ 00:29:05.318 { 00:29:05.318 "dma_device_id": "system", 00:29:05.318 "dma_device_type": 1 00:29:05.318 }, 00:29:05.318 { 00:29:05.318 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:05.318 "dma_device_type": 2 00:29:05.318 }, 00:29:05.318 { 00:29:05.318 "dma_device_id": "system", 00:29:05.318 "dma_device_type": 1 00:29:05.318 }, 00:29:05.318 { 00:29:05.318 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:05.318 "dma_device_type": 2 00:29:05.318 }, 00:29:05.318 { 00:29:05.318 "dma_device_id": "system", 00:29:05.318 "dma_device_type": 1 00:29:05.318 }, 00:29:05.318 { 00:29:05.318 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:05.318 "dma_device_type": 2 00:29:05.318 }, 00:29:05.318 { 00:29:05.318 "dma_device_id": "system", 00:29:05.318 "dma_device_type": 1 00:29:05.318 }, 00:29:05.318 { 00:29:05.318 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:05.318 "dma_device_type": 2 00:29:05.318 } 00:29:05.318 ], 00:29:05.318 "driver_specific": { 00:29:05.318 "raid": { 00:29:05.318 "uuid": "c0e47f68-95ea-462a-be62-b9ed13b395f0", 00:29:05.318 "strip_size_kb": 64, 00:29:05.318 "state": "online", 00:29:05.318 "raid_level": "raid0", 00:29:05.318 "superblock": true, 00:29:05.318 "num_base_bdevs": 4, 00:29:05.318 "num_base_bdevs_discovered": 4, 00:29:05.318 "num_base_bdevs_operational": 4, 00:29:05.318 "base_bdevs_list": [ 00:29:05.318 { 00:29:05.318 "name": "BaseBdev1", 00:29:05.318 "uuid": "2c3a63e0-ccad-48fa-9aa8-c1e599609c3d", 00:29:05.318 "is_configured": true, 00:29:05.318 "data_offset": 2048, 00:29:05.318 "data_size": 63488 00:29:05.318 }, 00:29:05.318 { 00:29:05.318 "name": "BaseBdev2", 00:29:05.318 "uuid": "448e24d4-4b20-4e40-b2b6-adb97d216e65", 00:29:05.318 "is_configured": true, 00:29:05.318 "data_offset": 2048, 00:29:05.318 "data_size": 63488 00:29:05.318 }, 00:29:05.318 { 00:29:05.318 "name": "BaseBdev3", 00:29:05.318 "uuid": "232214f1-0d72-47b0-9514-29cff39e288c", 00:29:05.318 "is_configured": true, 00:29:05.318 "data_offset": 2048, 00:29:05.318 "data_size": 63488 00:29:05.318 }, 00:29:05.318 { 00:29:05.318 "name": "BaseBdev4", 00:29:05.318 "uuid": "3d08640e-ebf3-4102-8c46-92d933ee30af", 00:29:05.318 "is_configured": true, 00:29:05.318 "data_offset": 2048, 00:29:05.318 "data_size": 63488 00:29:05.318 } 00:29:05.318 ] 00:29:05.318 } 00:29:05.318 } 00:29:05.318 }' 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:29:05.318 BaseBdev2 00:29:05.318 BaseBdev3 00:29:05.318 BaseBdev4' 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:05.318 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:05.577 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:05.577 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:05.577 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:05.577 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:29:05.577 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:05.577 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:05.577 [2024-11-20 13:49:12.406673] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:29:05.577 [2024-11-20 13:49:12.406705] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:29:05.577 [2024-11-20 13:49:12.406761] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:05.577 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:05.577 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:29:05.577 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:29:05.577 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:29:05.577 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:29:05.577 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:29:05.577 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 3 00:29:05.577 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:05.577 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:29:05.577 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:29:05.577 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:05.577 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:05.577 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:05.577 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:05.577 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:05.577 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:05.577 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:05.577 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:05.577 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:05.577 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:05.578 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:05.578 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:05.578 "name": "Existed_Raid", 00:29:05.578 "uuid": "c0e47f68-95ea-462a-be62-b9ed13b395f0", 00:29:05.578 "strip_size_kb": 64, 00:29:05.578 "state": "offline", 00:29:05.578 "raid_level": "raid0", 00:29:05.578 "superblock": true, 00:29:05.578 "num_base_bdevs": 4, 00:29:05.578 "num_base_bdevs_discovered": 3, 00:29:05.578 "num_base_bdevs_operational": 3, 00:29:05.578 "base_bdevs_list": [ 00:29:05.578 { 00:29:05.578 "name": null, 00:29:05.578 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:05.578 "is_configured": false, 00:29:05.578 "data_offset": 0, 00:29:05.578 "data_size": 63488 00:29:05.578 }, 00:29:05.578 { 00:29:05.578 "name": "BaseBdev2", 00:29:05.578 "uuid": "448e24d4-4b20-4e40-b2b6-adb97d216e65", 00:29:05.578 "is_configured": true, 00:29:05.578 "data_offset": 2048, 00:29:05.578 "data_size": 63488 00:29:05.578 }, 00:29:05.578 { 00:29:05.578 "name": "BaseBdev3", 00:29:05.578 "uuid": "232214f1-0d72-47b0-9514-29cff39e288c", 00:29:05.578 "is_configured": true, 00:29:05.578 "data_offset": 2048, 00:29:05.578 "data_size": 63488 00:29:05.578 }, 00:29:05.578 { 00:29:05.578 "name": "BaseBdev4", 00:29:05.578 "uuid": "3d08640e-ebf3-4102-8c46-92d933ee30af", 00:29:05.578 "is_configured": true, 00:29:05.578 "data_offset": 2048, 00:29:05.578 "data_size": 63488 00:29:05.578 } 00:29:05.578 ] 00:29:05.578 }' 00:29:05.578 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:05.578 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:05.837 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:29:05.837 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:29:05.837 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:29:05.837 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:05.837 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:05.837 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:05.837 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:05.837 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:29:05.837 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:29:05.837 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:29:05.837 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:05.837 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:05.837 [2024-11-20 13:49:12.829003] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:29:05.837 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:05.837 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:29:05.837 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:29:06.095 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:29:06.095 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:06.095 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.095 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:06.095 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:06.095 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:29:06.095 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:29:06.095 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:29:06.095 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.095 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:06.095 [2024-11-20 13:49:12.927897] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:29:06.095 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:06.095 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:29:06.095 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:29:06.095 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:29:06.095 13:49:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:06.095 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.095 13:49:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:06.095 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:06.095 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:29:06.096 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:29:06.096 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:29:06.096 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.096 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:06.096 [2024-11-20 13:49:13.021455] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:29:06.096 [2024-11-20 13:49:13.021505] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:29:06.096 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:06.096 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:29:06.096 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:29:06.096 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:06.096 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.096 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:06.096 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:29:06.096 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:06.096 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:29:06.096 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:29:06.096 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:29:06.096 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:29:06.096 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:29:06.096 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:29:06.096 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.096 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:06.355 BaseBdev2 00:29:06.355 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:06.355 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:29:06.355 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:29:06.355 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:06.355 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:29:06.355 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:06.355 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:06.355 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:06.356 [ 00:29:06.356 { 00:29:06.356 "name": "BaseBdev2", 00:29:06.356 "aliases": [ 00:29:06.356 "76dee79c-c0b1-49ac-be31-bb27150691d9" 00:29:06.356 ], 00:29:06.356 "product_name": "Malloc disk", 00:29:06.356 "block_size": 512, 00:29:06.356 "num_blocks": 65536, 00:29:06.356 "uuid": "76dee79c-c0b1-49ac-be31-bb27150691d9", 00:29:06.356 "assigned_rate_limits": { 00:29:06.356 "rw_ios_per_sec": 0, 00:29:06.356 "rw_mbytes_per_sec": 0, 00:29:06.356 "r_mbytes_per_sec": 0, 00:29:06.356 "w_mbytes_per_sec": 0 00:29:06.356 }, 00:29:06.356 "claimed": false, 00:29:06.356 "zoned": false, 00:29:06.356 "supported_io_types": { 00:29:06.356 "read": true, 00:29:06.356 "write": true, 00:29:06.356 "unmap": true, 00:29:06.356 "flush": true, 00:29:06.356 "reset": true, 00:29:06.356 "nvme_admin": false, 00:29:06.356 "nvme_io": false, 00:29:06.356 "nvme_io_md": false, 00:29:06.356 "write_zeroes": true, 00:29:06.356 "zcopy": true, 00:29:06.356 "get_zone_info": false, 00:29:06.356 "zone_management": false, 00:29:06.356 "zone_append": false, 00:29:06.356 "compare": false, 00:29:06.356 "compare_and_write": false, 00:29:06.356 "abort": true, 00:29:06.356 "seek_hole": false, 00:29:06.356 "seek_data": false, 00:29:06.356 "copy": true, 00:29:06.356 "nvme_iov_md": false 00:29:06.356 }, 00:29:06.356 "memory_domains": [ 00:29:06.356 { 00:29:06.356 "dma_device_id": "system", 00:29:06.356 "dma_device_type": 1 00:29:06.356 }, 00:29:06.356 { 00:29:06.356 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:06.356 "dma_device_type": 2 00:29:06.356 } 00:29:06.356 ], 00:29:06.356 "driver_specific": {} 00:29:06.356 } 00:29:06.356 ] 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:06.356 BaseBdev3 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:06.356 [ 00:29:06.356 { 00:29:06.356 "name": "BaseBdev3", 00:29:06.356 "aliases": [ 00:29:06.356 "23bec484-f167-46a1-a511-ae496c0837ce" 00:29:06.356 ], 00:29:06.356 "product_name": "Malloc disk", 00:29:06.356 "block_size": 512, 00:29:06.356 "num_blocks": 65536, 00:29:06.356 "uuid": "23bec484-f167-46a1-a511-ae496c0837ce", 00:29:06.356 "assigned_rate_limits": { 00:29:06.356 "rw_ios_per_sec": 0, 00:29:06.356 "rw_mbytes_per_sec": 0, 00:29:06.356 "r_mbytes_per_sec": 0, 00:29:06.356 "w_mbytes_per_sec": 0 00:29:06.356 }, 00:29:06.356 "claimed": false, 00:29:06.356 "zoned": false, 00:29:06.356 "supported_io_types": { 00:29:06.356 "read": true, 00:29:06.356 "write": true, 00:29:06.356 "unmap": true, 00:29:06.356 "flush": true, 00:29:06.356 "reset": true, 00:29:06.356 "nvme_admin": false, 00:29:06.356 "nvme_io": false, 00:29:06.356 "nvme_io_md": false, 00:29:06.356 "write_zeroes": true, 00:29:06.356 "zcopy": true, 00:29:06.356 "get_zone_info": false, 00:29:06.356 "zone_management": false, 00:29:06.356 "zone_append": false, 00:29:06.356 "compare": false, 00:29:06.356 "compare_and_write": false, 00:29:06.356 "abort": true, 00:29:06.356 "seek_hole": false, 00:29:06.356 "seek_data": false, 00:29:06.356 "copy": true, 00:29:06.356 "nvme_iov_md": false 00:29:06.356 }, 00:29:06.356 "memory_domains": [ 00:29:06.356 { 00:29:06.356 "dma_device_id": "system", 00:29:06.356 "dma_device_type": 1 00:29:06.356 }, 00:29:06.356 { 00:29:06.356 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:06.356 "dma_device_type": 2 00:29:06.356 } 00:29:06.356 ], 00:29:06.356 "driver_specific": {} 00:29:06.356 } 00:29:06.356 ] 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:06.356 BaseBdev4 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.356 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:06.356 [ 00:29:06.356 { 00:29:06.356 "name": "BaseBdev4", 00:29:06.356 "aliases": [ 00:29:06.356 "9d2f6b07-fde4-4902-8188-2179438cffa8" 00:29:06.356 ], 00:29:06.356 "product_name": "Malloc disk", 00:29:06.356 "block_size": 512, 00:29:06.356 "num_blocks": 65536, 00:29:06.356 "uuid": "9d2f6b07-fde4-4902-8188-2179438cffa8", 00:29:06.356 "assigned_rate_limits": { 00:29:06.356 "rw_ios_per_sec": 0, 00:29:06.356 "rw_mbytes_per_sec": 0, 00:29:06.356 "r_mbytes_per_sec": 0, 00:29:06.356 "w_mbytes_per_sec": 0 00:29:06.356 }, 00:29:06.356 "claimed": false, 00:29:06.356 "zoned": false, 00:29:06.356 "supported_io_types": { 00:29:06.356 "read": true, 00:29:06.356 "write": true, 00:29:06.356 "unmap": true, 00:29:06.356 "flush": true, 00:29:06.356 "reset": true, 00:29:06.356 "nvme_admin": false, 00:29:06.356 "nvme_io": false, 00:29:06.356 "nvme_io_md": false, 00:29:06.356 "write_zeroes": true, 00:29:06.356 "zcopy": true, 00:29:06.356 "get_zone_info": false, 00:29:06.356 "zone_management": false, 00:29:06.356 "zone_append": false, 00:29:06.356 "compare": false, 00:29:06.356 "compare_and_write": false, 00:29:06.356 "abort": true, 00:29:06.356 "seek_hole": false, 00:29:06.356 "seek_data": false, 00:29:06.356 "copy": true, 00:29:06.356 "nvme_iov_md": false 00:29:06.356 }, 00:29:06.356 "memory_domains": [ 00:29:06.356 { 00:29:06.356 "dma_device_id": "system", 00:29:06.357 "dma_device_type": 1 00:29:06.357 }, 00:29:06.357 { 00:29:06.357 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:06.357 "dma_device_type": 2 00:29:06.357 } 00:29:06.357 ], 00:29:06.357 "driver_specific": {} 00:29:06.357 } 00:29:06.357 ] 00:29:06.357 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:06.357 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:29:06.357 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:29:06.357 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:29:06.357 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:29:06.357 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.357 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:06.357 [2024-11-20 13:49:13.283077] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:29:06.357 [2024-11-20 13:49:13.283128] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:29:06.357 [2024-11-20 13:49:13.283150] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:29:06.357 [2024-11-20 13:49:13.285069] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:29:06.357 [2024-11-20 13:49:13.285117] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:29:06.357 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:06.357 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:29:06.357 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:06.357 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:06.357 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:29:06.357 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:06.357 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:06.357 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:06.357 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:06.357 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:06.357 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:06.357 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:06.357 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:06.357 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.357 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:06.357 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:06.357 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:06.357 "name": "Existed_Raid", 00:29:06.357 "uuid": "e60799e1-3500-4092-aabe-41a45ba7c94d", 00:29:06.357 "strip_size_kb": 64, 00:29:06.357 "state": "configuring", 00:29:06.357 "raid_level": "raid0", 00:29:06.357 "superblock": true, 00:29:06.357 "num_base_bdevs": 4, 00:29:06.357 "num_base_bdevs_discovered": 3, 00:29:06.357 "num_base_bdevs_operational": 4, 00:29:06.357 "base_bdevs_list": [ 00:29:06.357 { 00:29:06.357 "name": "BaseBdev1", 00:29:06.357 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:06.357 "is_configured": false, 00:29:06.357 "data_offset": 0, 00:29:06.357 "data_size": 0 00:29:06.357 }, 00:29:06.357 { 00:29:06.357 "name": "BaseBdev2", 00:29:06.357 "uuid": "76dee79c-c0b1-49ac-be31-bb27150691d9", 00:29:06.357 "is_configured": true, 00:29:06.357 "data_offset": 2048, 00:29:06.357 "data_size": 63488 00:29:06.357 }, 00:29:06.357 { 00:29:06.357 "name": "BaseBdev3", 00:29:06.357 "uuid": "23bec484-f167-46a1-a511-ae496c0837ce", 00:29:06.357 "is_configured": true, 00:29:06.357 "data_offset": 2048, 00:29:06.357 "data_size": 63488 00:29:06.357 }, 00:29:06.357 { 00:29:06.357 "name": "BaseBdev4", 00:29:06.357 "uuid": "9d2f6b07-fde4-4902-8188-2179438cffa8", 00:29:06.357 "is_configured": true, 00:29:06.357 "data_offset": 2048, 00:29:06.357 "data_size": 63488 00:29:06.357 } 00:29:06.357 ] 00:29:06.357 }' 00:29:06.357 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:06.357 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:06.615 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:29:06.615 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.615 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:06.615 [2024-11-20 13:49:13.603173] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:29:06.615 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:06.615 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:29:06.615 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:06.615 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:06.615 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:29:06.615 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:06.615 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:06.615 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:06.615 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:06.615 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:06.615 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:06.615 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:06.615 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:06.615 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.615 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:06.615 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:06.615 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:06.615 "name": "Existed_Raid", 00:29:06.615 "uuid": "e60799e1-3500-4092-aabe-41a45ba7c94d", 00:29:06.615 "strip_size_kb": 64, 00:29:06.615 "state": "configuring", 00:29:06.615 "raid_level": "raid0", 00:29:06.615 "superblock": true, 00:29:06.615 "num_base_bdevs": 4, 00:29:06.615 "num_base_bdevs_discovered": 2, 00:29:06.615 "num_base_bdevs_operational": 4, 00:29:06.615 "base_bdevs_list": [ 00:29:06.615 { 00:29:06.615 "name": "BaseBdev1", 00:29:06.615 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:06.615 "is_configured": false, 00:29:06.615 "data_offset": 0, 00:29:06.615 "data_size": 0 00:29:06.615 }, 00:29:06.615 { 00:29:06.615 "name": null, 00:29:06.615 "uuid": "76dee79c-c0b1-49ac-be31-bb27150691d9", 00:29:06.615 "is_configured": false, 00:29:06.615 "data_offset": 0, 00:29:06.615 "data_size": 63488 00:29:06.615 }, 00:29:06.615 { 00:29:06.615 "name": "BaseBdev3", 00:29:06.615 "uuid": "23bec484-f167-46a1-a511-ae496c0837ce", 00:29:06.615 "is_configured": true, 00:29:06.615 "data_offset": 2048, 00:29:06.615 "data_size": 63488 00:29:06.615 }, 00:29:06.615 { 00:29:06.615 "name": "BaseBdev4", 00:29:06.615 "uuid": "9d2f6b07-fde4-4902-8188-2179438cffa8", 00:29:06.615 "is_configured": true, 00:29:06.615 "data_offset": 2048, 00:29:06.615 "data_size": 63488 00:29:06.615 } 00:29:06.615 ] 00:29:06.615 }' 00:29:06.615 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:06.615 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:06.874 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:06.874 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:06.874 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:06.874 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:29:06.874 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:07.132 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:29:07.132 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:29:07.132 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:07.132 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:07.132 [2024-11-20 13:49:13.979726] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:29:07.132 BaseBdev1 00:29:07.132 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:07.132 13:49:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:29:07.132 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:29:07.132 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:07.132 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:29:07.132 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:07.132 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:07.132 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:07.132 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:07.132 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:07.132 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:07.132 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:29:07.133 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:07.133 13:49:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:07.133 [ 00:29:07.133 { 00:29:07.133 "name": "BaseBdev1", 00:29:07.133 "aliases": [ 00:29:07.133 "b671a646-49cf-4d50-a8c1-f00905025611" 00:29:07.133 ], 00:29:07.133 "product_name": "Malloc disk", 00:29:07.133 "block_size": 512, 00:29:07.133 "num_blocks": 65536, 00:29:07.133 "uuid": "b671a646-49cf-4d50-a8c1-f00905025611", 00:29:07.133 "assigned_rate_limits": { 00:29:07.133 "rw_ios_per_sec": 0, 00:29:07.133 "rw_mbytes_per_sec": 0, 00:29:07.133 "r_mbytes_per_sec": 0, 00:29:07.133 "w_mbytes_per_sec": 0 00:29:07.133 }, 00:29:07.133 "claimed": true, 00:29:07.133 "claim_type": "exclusive_write", 00:29:07.133 "zoned": false, 00:29:07.133 "supported_io_types": { 00:29:07.133 "read": true, 00:29:07.133 "write": true, 00:29:07.133 "unmap": true, 00:29:07.133 "flush": true, 00:29:07.133 "reset": true, 00:29:07.133 "nvme_admin": false, 00:29:07.133 "nvme_io": false, 00:29:07.133 "nvme_io_md": false, 00:29:07.133 "write_zeroes": true, 00:29:07.133 "zcopy": true, 00:29:07.133 "get_zone_info": false, 00:29:07.133 "zone_management": false, 00:29:07.133 "zone_append": false, 00:29:07.133 "compare": false, 00:29:07.133 "compare_and_write": false, 00:29:07.133 "abort": true, 00:29:07.133 "seek_hole": false, 00:29:07.133 "seek_data": false, 00:29:07.133 "copy": true, 00:29:07.133 "nvme_iov_md": false 00:29:07.133 }, 00:29:07.133 "memory_domains": [ 00:29:07.133 { 00:29:07.133 "dma_device_id": "system", 00:29:07.133 "dma_device_type": 1 00:29:07.133 }, 00:29:07.133 { 00:29:07.133 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:07.133 "dma_device_type": 2 00:29:07.133 } 00:29:07.133 ], 00:29:07.133 "driver_specific": {} 00:29:07.133 } 00:29:07.133 ] 00:29:07.133 13:49:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:07.133 13:49:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:29:07.133 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:29:07.133 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:07.133 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:07.133 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:29:07.133 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:07.133 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:07.133 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:07.133 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:07.133 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:07.133 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:07.133 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:07.133 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:07.133 13:49:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:07.133 13:49:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:07.133 13:49:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:07.133 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:07.133 "name": "Existed_Raid", 00:29:07.133 "uuid": "e60799e1-3500-4092-aabe-41a45ba7c94d", 00:29:07.133 "strip_size_kb": 64, 00:29:07.133 "state": "configuring", 00:29:07.133 "raid_level": "raid0", 00:29:07.133 "superblock": true, 00:29:07.133 "num_base_bdevs": 4, 00:29:07.133 "num_base_bdevs_discovered": 3, 00:29:07.133 "num_base_bdevs_operational": 4, 00:29:07.133 "base_bdevs_list": [ 00:29:07.133 { 00:29:07.133 "name": "BaseBdev1", 00:29:07.133 "uuid": "b671a646-49cf-4d50-a8c1-f00905025611", 00:29:07.133 "is_configured": true, 00:29:07.133 "data_offset": 2048, 00:29:07.133 "data_size": 63488 00:29:07.133 }, 00:29:07.133 { 00:29:07.133 "name": null, 00:29:07.133 "uuid": "76dee79c-c0b1-49ac-be31-bb27150691d9", 00:29:07.133 "is_configured": false, 00:29:07.133 "data_offset": 0, 00:29:07.133 "data_size": 63488 00:29:07.133 }, 00:29:07.133 { 00:29:07.133 "name": "BaseBdev3", 00:29:07.133 "uuid": "23bec484-f167-46a1-a511-ae496c0837ce", 00:29:07.133 "is_configured": true, 00:29:07.133 "data_offset": 2048, 00:29:07.133 "data_size": 63488 00:29:07.133 }, 00:29:07.133 { 00:29:07.133 "name": "BaseBdev4", 00:29:07.133 "uuid": "9d2f6b07-fde4-4902-8188-2179438cffa8", 00:29:07.133 "is_configured": true, 00:29:07.133 "data_offset": 2048, 00:29:07.133 "data_size": 63488 00:29:07.133 } 00:29:07.133 ] 00:29:07.133 }' 00:29:07.133 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:07.133 13:49:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:07.391 [2024-11-20 13:49:14.323895] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:07.391 "name": "Existed_Raid", 00:29:07.391 "uuid": "e60799e1-3500-4092-aabe-41a45ba7c94d", 00:29:07.391 "strip_size_kb": 64, 00:29:07.391 "state": "configuring", 00:29:07.391 "raid_level": "raid0", 00:29:07.391 "superblock": true, 00:29:07.391 "num_base_bdevs": 4, 00:29:07.391 "num_base_bdevs_discovered": 2, 00:29:07.391 "num_base_bdevs_operational": 4, 00:29:07.391 "base_bdevs_list": [ 00:29:07.391 { 00:29:07.391 "name": "BaseBdev1", 00:29:07.391 "uuid": "b671a646-49cf-4d50-a8c1-f00905025611", 00:29:07.391 "is_configured": true, 00:29:07.391 "data_offset": 2048, 00:29:07.391 "data_size": 63488 00:29:07.391 }, 00:29:07.391 { 00:29:07.391 "name": null, 00:29:07.391 "uuid": "76dee79c-c0b1-49ac-be31-bb27150691d9", 00:29:07.391 "is_configured": false, 00:29:07.391 "data_offset": 0, 00:29:07.391 "data_size": 63488 00:29:07.391 }, 00:29:07.391 { 00:29:07.391 "name": null, 00:29:07.391 "uuid": "23bec484-f167-46a1-a511-ae496c0837ce", 00:29:07.391 "is_configured": false, 00:29:07.391 "data_offset": 0, 00:29:07.391 "data_size": 63488 00:29:07.391 }, 00:29:07.391 { 00:29:07.391 "name": "BaseBdev4", 00:29:07.391 "uuid": "9d2f6b07-fde4-4902-8188-2179438cffa8", 00:29:07.391 "is_configured": true, 00:29:07.391 "data_offset": 2048, 00:29:07.391 "data_size": 63488 00:29:07.391 } 00:29:07.391 ] 00:29:07.391 }' 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:07.391 13:49:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:07.650 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:07.650 13:49:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:07.650 13:49:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:07.650 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:29:07.650 13:49:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:07.650 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:29:07.650 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:29:07.650 13:49:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:07.650 13:49:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:07.650 [2024-11-20 13:49:14.683995] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:29:07.650 13:49:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:07.650 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:29:07.650 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:07.650 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:07.650 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:29:07.650 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:07.650 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:07.650 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:07.650 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:07.650 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:07.650 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:07.650 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:07.650 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:07.650 13:49:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:07.650 13:49:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:07.650 13:49:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:07.907 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:07.907 "name": "Existed_Raid", 00:29:07.907 "uuid": "e60799e1-3500-4092-aabe-41a45ba7c94d", 00:29:07.907 "strip_size_kb": 64, 00:29:07.907 "state": "configuring", 00:29:07.907 "raid_level": "raid0", 00:29:07.907 "superblock": true, 00:29:07.907 "num_base_bdevs": 4, 00:29:07.907 "num_base_bdevs_discovered": 3, 00:29:07.907 "num_base_bdevs_operational": 4, 00:29:07.907 "base_bdevs_list": [ 00:29:07.907 { 00:29:07.907 "name": "BaseBdev1", 00:29:07.907 "uuid": "b671a646-49cf-4d50-a8c1-f00905025611", 00:29:07.907 "is_configured": true, 00:29:07.907 "data_offset": 2048, 00:29:07.907 "data_size": 63488 00:29:07.907 }, 00:29:07.907 { 00:29:07.907 "name": null, 00:29:07.907 "uuid": "76dee79c-c0b1-49ac-be31-bb27150691d9", 00:29:07.907 "is_configured": false, 00:29:07.907 "data_offset": 0, 00:29:07.907 "data_size": 63488 00:29:07.907 }, 00:29:07.907 { 00:29:07.907 "name": "BaseBdev3", 00:29:07.907 "uuid": "23bec484-f167-46a1-a511-ae496c0837ce", 00:29:07.907 "is_configured": true, 00:29:07.907 "data_offset": 2048, 00:29:07.907 "data_size": 63488 00:29:07.907 }, 00:29:07.907 { 00:29:07.907 "name": "BaseBdev4", 00:29:07.907 "uuid": "9d2f6b07-fde4-4902-8188-2179438cffa8", 00:29:07.907 "is_configured": true, 00:29:07.907 "data_offset": 2048, 00:29:07.907 "data_size": 63488 00:29:07.907 } 00:29:07.907 ] 00:29:07.907 }' 00:29:07.907 13:49:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:07.907 13:49:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:08.166 [2024-11-20 13:49:15.044119] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:08.166 "name": "Existed_Raid", 00:29:08.166 "uuid": "e60799e1-3500-4092-aabe-41a45ba7c94d", 00:29:08.166 "strip_size_kb": 64, 00:29:08.166 "state": "configuring", 00:29:08.166 "raid_level": "raid0", 00:29:08.166 "superblock": true, 00:29:08.166 "num_base_bdevs": 4, 00:29:08.166 "num_base_bdevs_discovered": 2, 00:29:08.166 "num_base_bdevs_operational": 4, 00:29:08.166 "base_bdevs_list": [ 00:29:08.166 { 00:29:08.166 "name": null, 00:29:08.166 "uuid": "b671a646-49cf-4d50-a8c1-f00905025611", 00:29:08.166 "is_configured": false, 00:29:08.166 "data_offset": 0, 00:29:08.166 "data_size": 63488 00:29:08.166 }, 00:29:08.166 { 00:29:08.166 "name": null, 00:29:08.166 "uuid": "76dee79c-c0b1-49ac-be31-bb27150691d9", 00:29:08.166 "is_configured": false, 00:29:08.166 "data_offset": 0, 00:29:08.166 "data_size": 63488 00:29:08.166 }, 00:29:08.166 { 00:29:08.166 "name": "BaseBdev3", 00:29:08.166 "uuid": "23bec484-f167-46a1-a511-ae496c0837ce", 00:29:08.166 "is_configured": true, 00:29:08.166 "data_offset": 2048, 00:29:08.166 "data_size": 63488 00:29:08.166 }, 00:29:08.166 { 00:29:08.166 "name": "BaseBdev4", 00:29:08.166 "uuid": "9d2f6b07-fde4-4902-8188-2179438cffa8", 00:29:08.166 "is_configured": true, 00:29:08.166 "data_offset": 2048, 00:29:08.166 "data_size": 63488 00:29:08.166 } 00:29:08.166 ] 00:29:08.166 }' 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:08.166 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:08.427 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:08.427 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:29:08.427 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:08.427 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:08.427 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:08.427 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:29:08.427 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:29:08.427 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:08.427 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:08.427 [2024-11-20 13:49:15.450297] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:29:08.427 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:08.427 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:29:08.427 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:08.427 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:08.427 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:29:08.427 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:08.427 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:08.427 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:08.427 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:08.427 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:08.427 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:08.427 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:08.427 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:08.427 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:08.427 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:08.427 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:08.692 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:08.692 "name": "Existed_Raid", 00:29:08.692 "uuid": "e60799e1-3500-4092-aabe-41a45ba7c94d", 00:29:08.692 "strip_size_kb": 64, 00:29:08.692 "state": "configuring", 00:29:08.692 "raid_level": "raid0", 00:29:08.692 "superblock": true, 00:29:08.692 "num_base_bdevs": 4, 00:29:08.692 "num_base_bdevs_discovered": 3, 00:29:08.692 "num_base_bdevs_operational": 4, 00:29:08.692 "base_bdevs_list": [ 00:29:08.692 { 00:29:08.693 "name": null, 00:29:08.693 "uuid": "b671a646-49cf-4d50-a8c1-f00905025611", 00:29:08.693 "is_configured": false, 00:29:08.693 "data_offset": 0, 00:29:08.693 "data_size": 63488 00:29:08.693 }, 00:29:08.693 { 00:29:08.693 "name": "BaseBdev2", 00:29:08.693 "uuid": "76dee79c-c0b1-49ac-be31-bb27150691d9", 00:29:08.693 "is_configured": true, 00:29:08.693 "data_offset": 2048, 00:29:08.693 "data_size": 63488 00:29:08.693 }, 00:29:08.693 { 00:29:08.693 "name": "BaseBdev3", 00:29:08.693 "uuid": "23bec484-f167-46a1-a511-ae496c0837ce", 00:29:08.693 "is_configured": true, 00:29:08.693 "data_offset": 2048, 00:29:08.693 "data_size": 63488 00:29:08.693 }, 00:29:08.693 { 00:29:08.693 "name": "BaseBdev4", 00:29:08.693 "uuid": "9d2f6b07-fde4-4902-8188-2179438cffa8", 00:29:08.693 "is_configured": true, 00:29:08.693 "data_offset": 2048, 00:29:08.693 "data_size": 63488 00:29:08.693 } 00:29:08.693 ] 00:29:08.693 }' 00:29:08.693 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:08.693 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u b671a646-49cf-4d50-a8c1-f00905025611 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:08.951 NewBaseBdev 00:29:08.951 [2024-11-20 13:49:15.855361] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:29:08.951 [2024-11-20 13:49:15.855602] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:29:08.951 [2024-11-20 13:49:15.855613] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:29:08.951 [2024-11-20 13:49:15.855847] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:29:08.951 [2024-11-20 13:49:15.855959] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:29:08.951 [2024-11-20 13:49:15.855968] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:29:08.951 [2024-11-20 13:49:15.856109] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:08.951 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:08.951 [ 00:29:08.951 { 00:29:08.951 "name": "NewBaseBdev", 00:29:08.951 "aliases": [ 00:29:08.951 "b671a646-49cf-4d50-a8c1-f00905025611" 00:29:08.951 ], 00:29:08.951 "product_name": "Malloc disk", 00:29:08.951 "block_size": 512, 00:29:08.951 "num_blocks": 65536, 00:29:08.951 "uuid": "b671a646-49cf-4d50-a8c1-f00905025611", 00:29:08.951 "assigned_rate_limits": { 00:29:08.951 "rw_ios_per_sec": 0, 00:29:08.951 "rw_mbytes_per_sec": 0, 00:29:08.951 "r_mbytes_per_sec": 0, 00:29:08.951 "w_mbytes_per_sec": 0 00:29:08.951 }, 00:29:08.951 "claimed": true, 00:29:08.951 "claim_type": "exclusive_write", 00:29:08.951 "zoned": false, 00:29:08.951 "supported_io_types": { 00:29:08.951 "read": true, 00:29:08.951 "write": true, 00:29:08.951 "unmap": true, 00:29:08.951 "flush": true, 00:29:08.951 "reset": true, 00:29:08.951 "nvme_admin": false, 00:29:08.951 "nvme_io": false, 00:29:08.951 "nvme_io_md": false, 00:29:08.951 "write_zeroes": true, 00:29:08.951 "zcopy": true, 00:29:08.951 "get_zone_info": false, 00:29:08.951 "zone_management": false, 00:29:08.951 "zone_append": false, 00:29:08.951 "compare": false, 00:29:08.951 "compare_and_write": false, 00:29:08.952 "abort": true, 00:29:08.952 "seek_hole": false, 00:29:08.952 "seek_data": false, 00:29:08.952 "copy": true, 00:29:08.952 "nvme_iov_md": false 00:29:08.952 }, 00:29:08.952 "memory_domains": [ 00:29:08.952 { 00:29:08.952 "dma_device_id": "system", 00:29:08.952 "dma_device_type": 1 00:29:08.952 }, 00:29:08.952 { 00:29:08.952 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:08.952 "dma_device_type": 2 00:29:08.952 } 00:29:08.952 ], 00:29:08.952 "driver_specific": {} 00:29:08.952 } 00:29:08.952 ] 00:29:08.952 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:08.952 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:29:08.952 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:29:08.952 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:08.952 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:08.952 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:29:08.952 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:08.952 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:08.952 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:08.952 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:08.952 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:08.952 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:08.952 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:08.952 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:08.952 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:08.952 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:08.952 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:08.952 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:08.952 "name": "Existed_Raid", 00:29:08.952 "uuid": "e60799e1-3500-4092-aabe-41a45ba7c94d", 00:29:08.952 "strip_size_kb": 64, 00:29:08.952 "state": "online", 00:29:08.952 "raid_level": "raid0", 00:29:08.952 "superblock": true, 00:29:08.952 "num_base_bdevs": 4, 00:29:08.952 "num_base_bdevs_discovered": 4, 00:29:08.952 "num_base_bdevs_operational": 4, 00:29:08.952 "base_bdevs_list": [ 00:29:08.952 { 00:29:08.952 "name": "NewBaseBdev", 00:29:08.952 "uuid": "b671a646-49cf-4d50-a8c1-f00905025611", 00:29:08.952 "is_configured": true, 00:29:08.952 "data_offset": 2048, 00:29:08.952 "data_size": 63488 00:29:08.952 }, 00:29:08.952 { 00:29:08.952 "name": "BaseBdev2", 00:29:08.952 "uuid": "76dee79c-c0b1-49ac-be31-bb27150691d9", 00:29:08.952 "is_configured": true, 00:29:08.952 "data_offset": 2048, 00:29:08.952 "data_size": 63488 00:29:08.952 }, 00:29:08.952 { 00:29:08.952 "name": "BaseBdev3", 00:29:08.952 "uuid": "23bec484-f167-46a1-a511-ae496c0837ce", 00:29:08.952 "is_configured": true, 00:29:08.952 "data_offset": 2048, 00:29:08.952 "data_size": 63488 00:29:08.952 }, 00:29:08.952 { 00:29:08.952 "name": "BaseBdev4", 00:29:08.952 "uuid": "9d2f6b07-fde4-4902-8188-2179438cffa8", 00:29:08.952 "is_configured": true, 00:29:08.952 "data_offset": 2048, 00:29:08.952 "data_size": 63488 00:29:08.952 } 00:29:08.952 ] 00:29:08.952 }' 00:29:08.952 13:49:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:08.952 13:49:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:09.210 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:29:09.210 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:29:09.210 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:29:09.210 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:29:09.210 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:29:09.210 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:29:09.210 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:29:09.210 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:29:09.210 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:09.210 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:09.210 [2024-11-20 13:49:16.203814] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:09.210 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:09.210 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:09.210 "name": "Existed_Raid", 00:29:09.210 "aliases": [ 00:29:09.210 "e60799e1-3500-4092-aabe-41a45ba7c94d" 00:29:09.210 ], 00:29:09.210 "product_name": "Raid Volume", 00:29:09.210 "block_size": 512, 00:29:09.210 "num_blocks": 253952, 00:29:09.210 "uuid": "e60799e1-3500-4092-aabe-41a45ba7c94d", 00:29:09.210 "assigned_rate_limits": { 00:29:09.210 "rw_ios_per_sec": 0, 00:29:09.210 "rw_mbytes_per_sec": 0, 00:29:09.210 "r_mbytes_per_sec": 0, 00:29:09.210 "w_mbytes_per_sec": 0 00:29:09.210 }, 00:29:09.210 "claimed": false, 00:29:09.210 "zoned": false, 00:29:09.210 "supported_io_types": { 00:29:09.210 "read": true, 00:29:09.210 "write": true, 00:29:09.210 "unmap": true, 00:29:09.210 "flush": true, 00:29:09.210 "reset": true, 00:29:09.210 "nvme_admin": false, 00:29:09.210 "nvme_io": false, 00:29:09.210 "nvme_io_md": false, 00:29:09.210 "write_zeroes": true, 00:29:09.210 "zcopy": false, 00:29:09.210 "get_zone_info": false, 00:29:09.210 "zone_management": false, 00:29:09.210 "zone_append": false, 00:29:09.210 "compare": false, 00:29:09.210 "compare_and_write": false, 00:29:09.210 "abort": false, 00:29:09.210 "seek_hole": false, 00:29:09.210 "seek_data": false, 00:29:09.210 "copy": false, 00:29:09.210 "nvme_iov_md": false 00:29:09.210 }, 00:29:09.210 "memory_domains": [ 00:29:09.210 { 00:29:09.210 "dma_device_id": "system", 00:29:09.210 "dma_device_type": 1 00:29:09.210 }, 00:29:09.210 { 00:29:09.210 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:09.210 "dma_device_type": 2 00:29:09.210 }, 00:29:09.210 { 00:29:09.210 "dma_device_id": "system", 00:29:09.210 "dma_device_type": 1 00:29:09.210 }, 00:29:09.210 { 00:29:09.210 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:09.210 "dma_device_type": 2 00:29:09.210 }, 00:29:09.210 { 00:29:09.210 "dma_device_id": "system", 00:29:09.210 "dma_device_type": 1 00:29:09.210 }, 00:29:09.210 { 00:29:09.210 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:09.210 "dma_device_type": 2 00:29:09.210 }, 00:29:09.210 { 00:29:09.210 "dma_device_id": "system", 00:29:09.210 "dma_device_type": 1 00:29:09.210 }, 00:29:09.210 { 00:29:09.210 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:09.210 "dma_device_type": 2 00:29:09.210 } 00:29:09.210 ], 00:29:09.210 "driver_specific": { 00:29:09.210 "raid": { 00:29:09.210 "uuid": "e60799e1-3500-4092-aabe-41a45ba7c94d", 00:29:09.210 "strip_size_kb": 64, 00:29:09.210 "state": "online", 00:29:09.210 "raid_level": "raid0", 00:29:09.210 "superblock": true, 00:29:09.210 "num_base_bdevs": 4, 00:29:09.210 "num_base_bdevs_discovered": 4, 00:29:09.210 "num_base_bdevs_operational": 4, 00:29:09.210 "base_bdevs_list": [ 00:29:09.210 { 00:29:09.210 "name": "NewBaseBdev", 00:29:09.210 "uuid": "b671a646-49cf-4d50-a8c1-f00905025611", 00:29:09.210 "is_configured": true, 00:29:09.210 "data_offset": 2048, 00:29:09.210 "data_size": 63488 00:29:09.210 }, 00:29:09.210 { 00:29:09.210 "name": "BaseBdev2", 00:29:09.210 "uuid": "76dee79c-c0b1-49ac-be31-bb27150691d9", 00:29:09.210 "is_configured": true, 00:29:09.210 "data_offset": 2048, 00:29:09.210 "data_size": 63488 00:29:09.210 }, 00:29:09.210 { 00:29:09.210 "name": "BaseBdev3", 00:29:09.211 "uuid": "23bec484-f167-46a1-a511-ae496c0837ce", 00:29:09.211 "is_configured": true, 00:29:09.211 "data_offset": 2048, 00:29:09.211 "data_size": 63488 00:29:09.211 }, 00:29:09.211 { 00:29:09.211 "name": "BaseBdev4", 00:29:09.211 "uuid": "9d2f6b07-fde4-4902-8188-2179438cffa8", 00:29:09.211 "is_configured": true, 00:29:09.211 "data_offset": 2048, 00:29:09.211 "data_size": 63488 00:29:09.211 } 00:29:09.211 ] 00:29:09.211 } 00:29:09.211 } 00:29:09.211 }' 00:29:09.211 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:29:09.211 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:29:09.211 BaseBdev2 00:29:09.211 BaseBdev3 00:29:09.211 BaseBdev4' 00:29:09.211 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:09.469 [2024-11-20 13:49:16.423530] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:29:09.469 [2024-11-20 13:49:16.423563] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:29:09.469 [2024-11-20 13:49:16.423640] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:09.469 [2024-11-20 13:49:16.423710] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:29:09.469 [2024-11-20 13:49:16.423719] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 68253 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 68253 ']' 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 68253 00:29:09.469 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:29:09.470 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:29:09.470 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 68253 00:29:09.470 killing process with pid 68253 00:29:09.470 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:29:09.470 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:29:09.470 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 68253' 00:29:09.470 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 68253 00:29:09.470 [2024-11-20 13:49:16.453442] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:29:09.470 13:49:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 68253 00:29:09.727 [2024-11-20 13:49:16.662135] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:29:10.293 13:49:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:29:10.293 00:29:10.293 real 0m8.256s 00:29:10.293 user 0m13.066s 00:29:10.293 sys 0m1.514s 00:29:10.293 13:49:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:29:10.293 ************************************ 00:29:10.293 END TEST raid_state_function_test_sb 00:29:10.293 ************************************ 00:29:10.293 13:49:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:10.293 13:49:17 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 4 00:29:10.293 13:49:17 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:29:10.293 13:49:17 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:29:10.293 13:49:17 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:29:10.293 ************************************ 00:29:10.293 START TEST raid_superblock_test 00:29:10.293 ************************************ 00:29:10.293 13:49:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid0 4 00:29:10.293 13:49:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:29:10.293 13:49:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:29:10.293 13:49:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:29:10.293 13:49:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:29:10.293 13:49:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:29:10.293 13:49:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:29:10.293 13:49:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:29:10.293 13:49:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:29:10.293 13:49:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:29:10.293 13:49:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:29:10.293 13:49:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:29:10.293 13:49:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:29:10.293 13:49:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:29:10.293 13:49:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:29:10.293 13:49:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:29:10.293 13:49:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:29:10.293 13:49:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=68890 00:29:10.293 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:10.293 13:49:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 68890 00:29:10.293 13:49:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 68890 ']' 00:29:10.293 13:49:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:10.293 13:49:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:29:10.293 13:49:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:10.294 13:49:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:29:10.294 13:49:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:10.294 13:49:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:29:10.551 [2024-11-20 13:49:17.395236] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:29:10.551 [2024-11-20 13:49:17.395479] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid68890 ] 00:29:10.551 [2024-11-20 13:49:17.547653] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:10.809 [2024-11-20 13:49:17.652158] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:29:10.809 [2024-11-20 13:49:17.774291] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:10.809 [2024-11-20 13:49:17.774480] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:11.375 malloc1 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:11.375 [2024-11-20 13:49:18.367550] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:29:11.375 [2024-11-20 13:49:18.367733] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:11.375 [2024-11-20 13:49:18.367777] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:29:11.375 [2024-11-20 13:49:18.368131] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:11.375 [2024-11-20 13:49:18.370211] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:11.375 [2024-11-20 13:49:18.370318] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:29:11.375 pt1 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:11.375 malloc2 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:11.375 [2024-11-20 13:49:18.405140] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:29:11.375 [2024-11-20 13:49:18.405183] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:11.375 [2024-11-20 13:49:18.405206] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:29:11.375 [2024-11-20 13:49:18.405214] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:11.375 [2024-11-20 13:49:18.407115] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:11.375 [2024-11-20 13:49:18.407142] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:29:11.375 pt2 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:11.375 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:11.634 malloc3 00:29:11.634 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:11.634 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:29:11.634 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:11.634 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:11.634 [2024-11-20 13:49:18.458746] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:29:11.634 [2024-11-20 13:49:18.458913] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:11.634 [2024-11-20 13:49:18.458941] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:29:11.634 [2024-11-20 13:49:18.458949] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:11.634 [2024-11-20 13:49:18.460904] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:11.634 [2024-11-20 13:49:18.460933] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:29:11.634 pt3 00:29:11.634 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:11.634 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:29:11.634 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:29:11.634 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:29:11.634 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:29:11.634 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:29:11.634 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:29:11.634 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:29:11.634 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:29:11.634 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:29:11.634 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:11.634 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:11.634 malloc4 00:29:11.634 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:11.634 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:29:11.634 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:11.634 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:11.634 [2024-11-20 13:49:18.492156] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:29:11.634 [2024-11-20 13:49:18.492197] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:11.635 [2024-11-20 13:49:18.492213] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:29:11.635 [2024-11-20 13:49:18.492221] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:11.635 [2024-11-20 13:49:18.494078] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:11.635 [2024-11-20 13:49:18.494105] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:29:11.635 pt4 00:29:11.635 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:11.635 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:29:11.635 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:29:11.635 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:29:11.635 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:11.635 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:11.635 [2024-11-20 13:49:18.500189] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:29:11.635 [2024-11-20 13:49:18.501767] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:29:11.635 [2024-11-20 13:49:18.501929] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:29:11.635 [2024-11-20 13:49:18.501973] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:29:11.635 [2024-11-20 13:49:18.502134] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:29:11.635 [2024-11-20 13:49:18.502143] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:29:11.635 [2024-11-20 13:49:18.502359] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:29:11.635 [2024-11-20 13:49:18.502480] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:29:11.635 [2024-11-20 13:49:18.502489] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:29:11.635 [2024-11-20 13:49:18.502598] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:11.635 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:11.635 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:29:11.635 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:11.635 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:11.635 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:29:11.635 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:11.635 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:11.635 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:11.635 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:11.635 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:11.635 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:11.635 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:11.635 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:11.635 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:11.635 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:11.635 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:11.635 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:11.635 "name": "raid_bdev1", 00:29:11.635 "uuid": "41a3d3a3-fd0e-419a-b46e-16d92f8ddfc7", 00:29:11.635 "strip_size_kb": 64, 00:29:11.635 "state": "online", 00:29:11.635 "raid_level": "raid0", 00:29:11.635 "superblock": true, 00:29:11.635 "num_base_bdevs": 4, 00:29:11.635 "num_base_bdevs_discovered": 4, 00:29:11.635 "num_base_bdevs_operational": 4, 00:29:11.635 "base_bdevs_list": [ 00:29:11.635 { 00:29:11.635 "name": "pt1", 00:29:11.635 "uuid": "00000000-0000-0000-0000-000000000001", 00:29:11.635 "is_configured": true, 00:29:11.635 "data_offset": 2048, 00:29:11.635 "data_size": 63488 00:29:11.635 }, 00:29:11.635 { 00:29:11.635 "name": "pt2", 00:29:11.635 "uuid": "00000000-0000-0000-0000-000000000002", 00:29:11.635 "is_configured": true, 00:29:11.635 "data_offset": 2048, 00:29:11.635 "data_size": 63488 00:29:11.635 }, 00:29:11.635 { 00:29:11.635 "name": "pt3", 00:29:11.635 "uuid": "00000000-0000-0000-0000-000000000003", 00:29:11.635 "is_configured": true, 00:29:11.635 "data_offset": 2048, 00:29:11.635 "data_size": 63488 00:29:11.635 }, 00:29:11.635 { 00:29:11.635 "name": "pt4", 00:29:11.635 "uuid": "00000000-0000-0000-0000-000000000004", 00:29:11.635 "is_configured": true, 00:29:11.635 "data_offset": 2048, 00:29:11.635 "data_size": 63488 00:29:11.635 } 00:29:11.635 ] 00:29:11.635 }' 00:29:11.635 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:11.635 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:11.895 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:29:11.895 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:29:11.895 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:29:11.895 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:29:11.895 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:29:11.895 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:29:11.895 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:29:11.895 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:29:11.895 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:11.895 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:11.895 [2024-11-20 13:49:18.832592] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:11.895 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:11.895 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:11.895 "name": "raid_bdev1", 00:29:11.895 "aliases": [ 00:29:11.895 "41a3d3a3-fd0e-419a-b46e-16d92f8ddfc7" 00:29:11.895 ], 00:29:11.895 "product_name": "Raid Volume", 00:29:11.895 "block_size": 512, 00:29:11.895 "num_blocks": 253952, 00:29:11.895 "uuid": "41a3d3a3-fd0e-419a-b46e-16d92f8ddfc7", 00:29:11.895 "assigned_rate_limits": { 00:29:11.895 "rw_ios_per_sec": 0, 00:29:11.895 "rw_mbytes_per_sec": 0, 00:29:11.895 "r_mbytes_per_sec": 0, 00:29:11.896 "w_mbytes_per_sec": 0 00:29:11.896 }, 00:29:11.896 "claimed": false, 00:29:11.896 "zoned": false, 00:29:11.896 "supported_io_types": { 00:29:11.896 "read": true, 00:29:11.896 "write": true, 00:29:11.896 "unmap": true, 00:29:11.896 "flush": true, 00:29:11.896 "reset": true, 00:29:11.896 "nvme_admin": false, 00:29:11.896 "nvme_io": false, 00:29:11.896 "nvme_io_md": false, 00:29:11.896 "write_zeroes": true, 00:29:11.896 "zcopy": false, 00:29:11.896 "get_zone_info": false, 00:29:11.896 "zone_management": false, 00:29:11.896 "zone_append": false, 00:29:11.896 "compare": false, 00:29:11.896 "compare_and_write": false, 00:29:11.896 "abort": false, 00:29:11.896 "seek_hole": false, 00:29:11.896 "seek_data": false, 00:29:11.896 "copy": false, 00:29:11.896 "nvme_iov_md": false 00:29:11.896 }, 00:29:11.896 "memory_domains": [ 00:29:11.896 { 00:29:11.896 "dma_device_id": "system", 00:29:11.896 "dma_device_type": 1 00:29:11.896 }, 00:29:11.896 { 00:29:11.896 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:11.896 "dma_device_type": 2 00:29:11.896 }, 00:29:11.896 { 00:29:11.896 "dma_device_id": "system", 00:29:11.896 "dma_device_type": 1 00:29:11.896 }, 00:29:11.896 { 00:29:11.896 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:11.896 "dma_device_type": 2 00:29:11.896 }, 00:29:11.896 { 00:29:11.896 "dma_device_id": "system", 00:29:11.896 "dma_device_type": 1 00:29:11.896 }, 00:29:11.896 { 00:29:11.896 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:11.896 "dma_device_type": 2 00:29:11.896 }, 00:29:11.896 { 00:29:11.896 "dma_device_id": "system", 00:29:11.896 "dma_device_type": 1 00:29:11.896 }, 00:29:11.896 { 00:29:11.896 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:11.896 "dma_device_type": 2 00:29:11.896 } 00:29:11.896 ], 00:29:11.896 "driver_specific": { 00:29:11.896 "raid": { 00:29:11.896 "uuid": "41a3d3a3-fd0e-419a-b46e-16d92f8ddfc7", 00:29:11.896 "strip_size_kb": 64, 00:29:11.896 "state": "online", 00:29:11.896 "raid_level": "raid0", 00:29:11.896 "superblock": true, 00:29:11.896 "num_base_bdevs": 4, 00:29:11.896 "num_base_bdevs_discovered": 4, 00:29:11.896 "num_base_bdevs_operational": 4, 00:29:11.896 "base_bdevs_list": [ 00:29:11.896 { 00:29:11.896 "name": "pt1", 00:29:11.896 "uuid": "00000000-0000-0000-0000-000000000001", 00:29:11.896 "is_configured": true, 00:29:11.896 "data_offset": 2048, 00:29:11.896 "data_size": 63488 00:29:11.896 }, 00:29:11.896 { 00:29:11.896 "name": "pt2", 00:29:11.896 "uuid": "00000000-0000-0000-0000-000000000002", 00:29:11.896 "is_configured": true, 00:29:11.896 "data_offset": 2048, 00:29:11.896 "data_size": 63488 00:29:11.896 }, 00:29:11.896 { 00:29:11.896 "name": "pt3", 00:29:11.896 "uuid": "00000000-0000-0000-0000-000000000003", 00:29:11.896 "is_configured": true, 00:29:11.896 "data_offset": 2048, 00:29:11.896 "data_size": 63488 00:29:11.896 }, 00:29:11.896 { 00:29:11.896 "name": "pt4", 00:29:11.896 "uuid": "00000000-0000-0000-0000-000000000004", 00:29:11.896 "is_configured": true, 00:29:11.896 "data_offset": 2048, 00:29:11.896 "data_size": 63488 00:29:11.896 } 00:29:11.896 ] 00:29:11.896 } 00:29:11.896 } 00:29:11.896 }' 00:29:11.896 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:29:11.896 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:29:11.896 pt2 00:29:11.896 pt3 00:29:11.896 pt4' 00:29:11.896 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:11.896 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:29:11.896 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:11.896 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:29:11.896 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:11.896 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:11.896 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:11.896 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:12.196 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:12.196 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:12.196 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:12.196 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:12.196 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:29:12.196 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:12.196 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:12.196 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:12.196 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:12.196 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:12.196 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:12.196 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:12.196 13:49:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:29:12.196 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:12.196 13:49:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:12.196 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:12.196 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:12.196 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:12.196 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:12.196 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:29:12.196 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:12.196 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:12.196 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:12.196 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:12.196 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:12.196 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:12.196 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:29:12.196 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:12.196 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:12.196 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:29:12.196 [2024-11-20 13:49:19.056608] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:12.196 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:12.196 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=41a3d3a3-fd0e-419a-b46e-16d92f8ddfc7 00:29:12.196 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 41a3d3a3-fd0e-419a-b46e-16d92f8ddfc7 ']' 00:29:12.196 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:29:12.196 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:12.196 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:12.196 [2024-11-20 13:49:19.088302] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:29:12.196 [2024-11-20 13:49:19.088327] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:29:12.196 [2024-11-20 13:49:19.088408] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:12.196 [2024-11-20 13:49:19.088480] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:29:12.196 [2024-11-20 13:49:19.088495] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:12.197 [2024-11-20 13:49:19.200352] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:29:12.197 [2024-11-20 13:49:19.202140] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:29:12.197 [2024-11-20 13:49:19.202185] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:29:12.197 [2024-11-20 13:49:19.202215] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:29:12.197 [2024-11-20 13:49:19.202260] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:29:12.197 [2024-11-20 13:49:19.202306] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:29:12.197 [2024-11-20 13:49:19.202323] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:29:12.197 [2024-11-20 13:49:19.202339] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:29:12.197 [2024-11-20 13:49:19.202350] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:29:12.197 [2024-11-20 13:49:19.202362] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:29:12.197 request: 00:29:12.197 { 00:29:12.197 "name": "raid_bdev1", 00:29:12.197 "raid_level": "raid0", 00:29:12.197 "base_bdevs": [ 00:29:12.197 "malloc1", 00:29:12.197 "malloc2", 00:29:12.197 "malloc3", 00:29:12.197 "malloc4" 00:29:12.197 ], 00:29:12.197 "strip_size_kb": 64, 00:29:12.197 "superblock": false, 00:29:12.197 "method": "bdev_raid_create", 00:29:12.197 "req_id": 1 00:29:12.197 } 00:29:12.197 Got JSON-RPC error response 00:29:12.197 response: 00:29:12.197 { 00:29:12.197 "code": -17, 00:29:12.197 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:29:12.197 } 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:29:12.197 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:12.465 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:12.465 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:29:12.465 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:29:12.465 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:29:12.465 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:12.465 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:12.465 [2024-11-20 13:49:19.240354] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:29:12.465 [2024-11-20 13:49:19.240433] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:12.465 [2024-11-20 13:49:19.240456] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:29:12.465 [2024-11-20 13:49:19.240467] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:12.465 [2024-11-20 13:49:19.242550] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:12.465 [2024-11-20 13:49:19.242586] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:29:12.465 [2024-11-20 13:49:19.242667] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:29:12.465 [2024-11-20 13:49:19.242717] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:29:12.465 pt1 00:29:12.465 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:12.465 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 4 00:29:12.465 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:12.465 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:12.465 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:29:12.465 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:12.465 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:12.465 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:12.465 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:12.465 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:12.465 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:12.465 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:12.465 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:12.465 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:12.465 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:12.465 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:12.465 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:12.465 "name": "raid_bdev1", 00:29:12.465 "uuid": "41a3d3a3-fd0e-419a-b46e-16d92f8ddfc7", 00:29:12.465 "strip_size_kb": 64, 00:29:12.465 "state": "configuring", 00:29:12.465 "raid_level": "raid0", 00:29:12.465 "superblock": true, 00:29:12.465 "num_base_bdevs": 4, 00:29:12.465 "num_base_bdevs_discovered": 1, 00:29:12.465 "num_base_bdevs_operational": 4, 00:29:12.465 "base_bdevs_list": [ 00:29:12.465 { 00:29:12.465 "name": "pt1", 00:29:12.465 "uuid": "00000000-0000-0000-0000-000000000001", 00:29:12.465 "is_configured": true, 00:29:12.465 "data_offset": 2048, 00:29:12.465 "data_size": 63488 00:29:12.465 }, 00:29:12.465 { 00:29:12.465 "name": null, 00:29:12.465 "uuid": "00000000-0000-0000-0000-000000000002", 00:29:12.465 "is_configured": false, 00:29:12.465 "data_offset": 2048, 00:29:12.465 "data_size": 63488 00:29:12.465 }, 00:29:12.465 { 00:29:12.465 "name": null, 00:29:12.465 "uuid": "00000000-0000-0000-0000-000000000003", 00:29:12.465 "is_configured": false, 00:29:12.465 "data_offset": 2048, 00:29:12.466 "data_size": 63488 00:29:12.466 }, 00:29:12.466 { 00:29:12.466 "name": null, 00:29:12.466 "uuid": "00000000-0000-0000-0000-000000000004", 00:29:12.466 "is_configured": false, 00:29:12.466 "data_offset": 2048, 00:29:12.466 "data_size": 63488 00:29:12.466 } 00:29:12.466 ] 00:29:12.466 }' 00:29:12.466 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:12.466 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:12.724 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:29:12.724 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:29:12.724 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:12.724 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:12.724 [2024-11-20 13:49:19.564427] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:29:12.724 [2024-11-20 13:49:19.564506] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:12.724 [2024-11-20 13:49:19.564527] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:29:12.724 [2024-11-20 13:49:19.564537] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:12.724 [2024-11-20 13:49:19.564956] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:12.724 [2024-11-20 13:49:19.564970] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:29:12.724 [2024-11-20 13:49:19.565052] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:29:12.724 [2024-11-20 13:49:19.565074] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:29:12.724 pt2 00:29:12.724 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:12.724 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:29:12.724 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:12.724 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:12.724 [2024-11-20 13:49:19.572441] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:29:12.724 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:12.724 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 4 00:29:12.724 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:12.724 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:12.724 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:29:12.724 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:12.724 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:12.724 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:12.724 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:12.724 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:12.724 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:12.724 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:12.724 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:12.724 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:12.724 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:12.724 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:12.724 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:12.724 "name": "raid_bdev1", 00:29:12.724 "uuid": "41a3d3a3-fd0e-419a-b46e-16d92f8ddfc7", 00:29:12.724 "strip_size_kb": 64, 00:29:12.724 "state": "configuring", 00:29:12.724 "raid_level": "raid0", 00:29:12.724 "superblock": true, 00:29:12.724 "num_base_bdevs": 4, 00:29:12.724 "num_base_bdevs_discovered": 1, 00:29:12.724 "num_base_bdevs_operational": 4, 00:29:12.724 "base_bdevs_list": [ 00:29:12.724 { 00:29:12.724 "name": "pt1", 00:29:12.724 "uuid": "00000000-0000-0000-0000-000000000001", 00:29:12.724 "is_configured": true, 00:29:12.724 "data_offset": 2048, 00:29:12.724 "data_size": 63488 00:29:12.724 }, 00:29:12.724 { 00:29:12.724 "name": null, 00:29:12.724 "uuid": "00000000-0000-0000-0000-000000000002", 00:29:12.724 "is_configured": false, 00:29:12.724 "data_offset": 0, 00:29:12.724 "data_size": 63488 00:29:12.724 }, 00:29:12.724 { 00:29:12.724 "name": null, 00:29:12.724 "uuid": "00000000-0000-0000-0000-000000000003", 00:29:12.724 "is_configured": false, 00:29:12.724 "data_offset": 2048, 00:29:12.724 "data_size": 63488 00:29:12.724 }, 00:29:12.724 { 00:29:12.724 "name": null, 00:29:12.724 "uuid": "00000000-0000-0000-0000-000000000004", 00:29:12.724 "is_configured": false, 00:29:12.724 "data_offset": 2048, 00:29:12.724 "data_size": 63488 00:29:12.725 } 00:29:12.725 ] 00:29:12.725 }' 00:29:12.725 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:12.725 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:12.983 [2024-11-20 13:49:19.884478] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:29:12.983 [2024-11-20 13:49:19.884547] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:12.983 [2024-11-20 13:49:19.884567] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:29:12.983 [2024-11-20 13:49:19.884575] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:12.983 [2024-11-20 13:49:19.884996] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:12.983 [2024-11-20 13:49:19.885008] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:29:12.983 [2024-11-20 13:49:19.885083] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:29:12.983 [2024-11-20 13:49:19.885102] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:29:12.983 pt2 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:12.983 [2024-11-20 13:49:19.896463] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:29:12.983 [2024-11-20 13:49:19.896516] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:12.983 [2024-11-20 13:49:19.896536] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:29:12.983 [2024-11-20 13:49:19.896545] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:12.983 [2024-11-20 13:49:19.896930] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:12.983 [2024-11-20 13:49:19.896946] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:29:12.983 [2024-11-20 13:49:19.897022] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:29:12.983 [2024-11-20 13:49:19.897042] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:29:12.983 pt3 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:12.983 [2024-11-20 13:49:19.904439] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:29:12.983 [2024-11-20 13:49:19.904484] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:12.983 [2024-11-20 13:49:19.904503] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:29:12.983 [2024-11-20 13:49:19.904510] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:12.983 [2024-11-20 13:49:19.904875] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:12.983 [2024-11-20 13:49:19.904890] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:29:12.983 [2024-11-20 13:49:19.904951] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:29:12.983 [2024-11-20 13:49:19.904973] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:29:12.983 [2024-11-20 13:49:19.905106] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:29:12.983 [2024-11-20 13:49:19.905120] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:29:12.983 [2024-11-20 13:49:19.905346] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:29:12.983 [2024-11-20 13:49:19.905458] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:29:12.983 [2024-11-20 13:49:19.905467] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:29:12.983 [2024-11-20 13:49:19.905575] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:12.983 pt4 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:12.983 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:12.984 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:12.984 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:12.984 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:12.984 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:12.984 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:12.984 "name": "raid_bdev1", 00:29:12.984 "uuid": "41a3d3a3-fd0e-419a-b46e-16d92f8ddfc7", 00:29:12.984 "strip_size_kb": 64, 00:29:12.984 "state": "online", 00:29:12.984 "raid_level": "raid0", 00:29:12.984 "superblock": true, 00:29:12.984 "num_base_bdevs": 4, 00:29:12.984 "num_base_bdevs_discovered": 4, 00:29:12.984 "num_base_bdevs_operational": 4, 00:29:12.984 "base_bdevs_list": [ 00:29:12.984 { 00:29:12.984 "name": "pt1", 00:29:12.984 "uuid": "00000000-0000-0000-0000-000000000001", 00:29:12.984 "is_configured": true, 00:29:12.984 "data_offset": 2048, 00:29:12.984 "data_size": 63488 00:29:12.984 }, 00:29:12.984 { 00:29:12.984 "name": "pt2", 00:29:12.984 "uuid": "00000000-0000-0000-0000-000000000002", 00:29:12.984 "is_configured": true, 00:29:12.984 "data_offset": 2048, 00:29:12.984 "data_size": 63488 00:29:12.984 }, 00:29:12.984 { 00:29:12.984 "name": "pt3", 00:29:12.984 "uuid": "00000000-0000-0000-0000-000000000003", 00:29:12.984 "is_configured": true, 00:29:12.984 "data_offset": 2048, 00:29:12.984 "data_size": 63488 00:29:12.984 }, 00:29:12.984 { 00:29:12.984 "name": "pt4", 00:29:12.984 "uuid": "00000000-0000-0000-0000-000000000004", 00:29:12.984 "is_configured": true, 00:29:12.984 "data_offset": 2048, 00:29:12.984 "data_size": 63488 00:29:12.984 } 00:29:12.984 ] 00:29:12.984 }' 00:29:12.984 13:49:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:12.984 13:49:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:13.242 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:29:13.242 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:29:13.242 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:29:13.242 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:29:13.242 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:29:13.242 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:29:13.242 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:29:13.242 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:29:13.242 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:13.243 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:13.243 [2024-11-20 13:49:20.224864] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:13.243 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:13.243 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:13.243 "name": "raid_bdev1", 00:29:13.243 "aliases": [ 00:29:13.243 "41a3d3a3-fd0e-419a-b46e-16d92f8ddfc7" 00:29:13.243 ], 00:29:13.243 "product_name": "Raid Volume", 00:29:13.243 "block_size": 512, 00:29:13.243 "num_blocks": 253952, 00:29:13.243 "uuid": "41a3d3a3-fd0e-419a-b46e-16d92f8ddfc7", 00:29:13.243 "assigned_rate_limits": { 00:29:13.243 "rw_ios_per_sec": 0, 00:29:13.243 "rw_mbytes_per_sec": 0, 00:29:13.243 "r_mbytes_per_sec": 0, 00:29:13.243 "w_mbytes_per_sec": 0 00:29:13.243 }, 00:29:13.243 "claimed": false, 00:29:13.243 "zoned": false, 00:29:13.243 "supported_io_types": { 00:29:13.243 "read": true, 00:29:13.243 "write": true, 00:29:13.243 "unmap": true, 00:29:13.243 "flush": true, 00:29:13.243 "reset": true, 00:29:13.243 "nvme_admin": false, 00:29:13.243 "nvme_io": false, 00:29:13.243 "nvme_io_md": false, 00:29:13.243 "write_zeroes": true, 00:29:13.243 "zcopy": false, 00:29:13.243 "get_zone_info": false, 00:29:13.243 "zone_management": false, 00:29:13.243 "zone_append": false, 00:29:13.243 "compare": false, 00:29:13.243 "compare_and_write": false, 00:29:13.243 "abort": false, 00:29:13.243 "seek_hole": false, 00:29:13.243 "seek_data": false, 00:29:13.243 "copy": false, 00:29:13.243 "nvme_iov_md": false 00:29:13.243 }, 00:29:13.243 "memory_domains": [ 00:29:13.243 { 00:29:13.243 "dma_device_id": "system", 00:29:13.243 "dma_device_type": 1 00:29:13.243 }, 00:29:13.243 { 00:29:13.243 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:13.243 "dma_device_type": 2 00:29:13.243 }, 00:29:13.243 { 00:29:13.243 "dma_device_id": "system", 00:29:13.243 "dma_device_type": 1 00:29:13.243 }, 00:29:13.243 { 00:29:13.243 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:13.243 "dma_device_type": 2 00:29:13.243 }, 00:29:13.243 { 00:29:13.243 "dma_device_id": "system", 00:29:13.243 "dma_device_type": 1 00:29:13.243 }, 00:29:13.243 { 00:29:13.243 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:13.243 "dma_device_type": 2 00:29:13.243 }, 00:29:13.243 { 00:29:13.243 "dma_device_id": "system", 00:29:13.243 "dma_device_type": 1 00:29:13.243 }, 00:29:13.243 { 00:29:13.243 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:13.243 "dma_device_type": 2 00:29:13.243 } 00:29:13.243 ], 00:29:13.243 "driver_specific": { 00:29:13.243 "raid": { 00:29:13.243 "uuid": "41a3d3a3-fd0e-419a-b46e-16d92f8ddfc7", 00:29:13.243 "strip_size_kb": 64, 00:29:13.243 "state": "online", 00:29:13.243 "raid_level": "raid0", 00:29:13.243 "superblock": true, 00:29:13.243 "num_base_bdevs": 4, 00:29:13.243 "num_base_bdevs_discovered": 4, 00:29:13.243 "num_base_bdevs_operational": 4, 00:29:13.243 "base_bdevs_list": [ 00:29:13.243 { 00:29:13.243 "name": "pt1", 00:29:13.243 "uuid": "00000000-0000-0000-0000-000000000001", 00:29:13.243 "is_configured": true, 00:29:13.243 "data_offset": 2048, 00:29:13.243 "data_size": 63488 00:29:13.243 }, 00:29:13.243 { 00:29:13.243 "name": "pt2", 00:29:13.243 "uuid": "00000000-0000-0000-0000-000000000002", 00:29:13.243 "is_configured": true, 00:29:13.243 "data_offset": 2048, 00:29:13.243 "data_size": 63488 00:29:13.243 }, 00:29:13.243 { 00:29:13.243 "name": "pt3", 00:29:13.243 "uuid": "00000000-0000-0000-0000-000000000003", 00:29:13.243 "is_configured": true, 00:29:13.243 "data_offset": 2048, 00:29:13.243 "data_size": 63488 00:29:13.243 }, 00:29:13.243 { 00:29:13.243 "name": "pt4", 00:29:13.243 "uuid": "00000000-0000-0000-0000-000000000004", 00:29:13.243 "is_configured": true, 00:29:13.243 "data_offset": 2048, 00:29:13.243 "data_size": 63488 00:29:13.243 } 00:29:13.243 ] 00:29:13.243 } 00:29:13.243 } 00:29:13.243 }' 00:29:13.243 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:29:13.243 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:29:13.243 pt2 00:29:13.243 pt3 00:29:13.243 pt4' 00:29:13.243 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:13.501 [2024-11-20 13:49:20.468877] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 41a3d3a3-fd0e-419a-b46e-16d92f8ddfc7 '!=' 41a3d3a3-fd0e-419a-b46e-16d92f8ddfc7 ']' 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 68890 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 68890 ']' 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 68890 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 68890 00:29:13.501 killing process with pid 68890 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 68890' 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 68890 00:29:13.501 [2024-11-20 13:49:20.518779] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:29:13.501 13:49:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 68890 00:29:13.501 [2024-11-20 13:49:20.518880] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:13.501 [2024-11-20 13:49:20.518961] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:29:13.501 [2024-11-20 13:49:20.518971] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:29:13.757 [2024-11-20 13:49:20.726352] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:29:14.320 13:49:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:29:14.320 00:29:14.320 real 0m4.017s 00:29:14.320 user 0m5.791s 00:29:14.320 sys 0m0.692s 00:29:14.320 13:49:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:29:14.320 13:49:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:14.320 ************************************ 00:29:14.320 END TEST raid_superblock_test 00:29:14.320 ************************************ 00:29:14.578 13:49:21 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 4 read 00:29:14.578 13:49:21 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:29:14.578 13:49:21 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:29:14.578 13:49:21 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:29:14.578 ************************************ 00:29:14.578 START TEST raid_read_error_test 00:29:14.578 ************************************ 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 4 read 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.UZKJbgzqP5 00:29:14.578 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=69138 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 69138 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 69138 ']' 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:29:14.578 13:49:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:14.578 [2024-11-20 13:49:21.478128] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:29:14.578 [2024-11-20 13:49:21.478269] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69138 ] 00:29:14.835 [2024-11-20 13:49:21.635291] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:14.835 [2024-11-20 13:49:21.737171] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:29:14.835 [2024-11-20 13:49:21.859913] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:14.835 [2024-11-20 13:49:21.859944] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:15.401 BaseBdev1_malloc 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:15.401 true 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:15.401 [2024-11-20 13:49:22.358348] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:29:15.401 [2024-11-20 13:49:22.358515] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:15.401 [2024-11-20 13:49:22.358541] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:29:15.401 [2024-11-20 13:49:22.358551] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:15.401 [2024-11-20 13:49:22.360480] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:15.401 [2024-11-20 13:49:22.360513] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:29:15.401 BaseBdev1 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:15.401 BaseBdev2_malloc 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:15.401 true 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:15.401 [2024-11-20 13:49:22.400639] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:29:15.401 [2024-11-20 13:49:22.400698] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:15.401 [2024-11-20 13:49:22.400713] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:29:15.401 [2024-11-20 13:49:22.400722] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:15.401 [2024-11-20 13:49:22.402569] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:15.401 [2024-11-20 13:49:22.402602] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:29:15.401 BaseBdev2 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:15.401 BaseBdev3_malloc 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:15.401 true 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:15.401 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:15.659 [2024-11-20 13:49:22.459261] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:29:15.659 [2024-11-20 13:49:22.459474] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:15.659 [2024-11-20 13:49:22.459497] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:29:15.659 [2024-11-20 13:49:22.459507] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:15.659 [2024-11-20 13:49:22.461399] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:15.659 [2024-11-20 13:49:22.461433] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:29:15.659 BaseBdev3 00:29:15.659 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:15.659 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:29:15.659 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:29:15.659 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:15.659 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:15.659 BaseBdev4_malloc 00:29:15.659 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:15.659 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:29:15.659 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:15.659 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:15.659 true 00:29:15.659 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:15.659 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:29:15.659 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:15.659 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:15.659 [2024-11-20 13:49:22.500819] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:29:15.659 [2024-11-20 13:49:22.500876] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:15.660 [2024-11-20 13:49:22.500894] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:29:15.660 [2024-11-20 13:49:22.500905] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:15.660 [2024-11-20 13:49:22.502814] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:15.660 [2024-11-20 13:49:22.502973] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:29:15.660 BaseBdev4 00:29:15.660 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:15.660 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:29:15.660 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:15.660 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:15.660 [2024-11-20 13:49:22.508887] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:29:15.660 [2024-11-20 13:49:22.510714] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:29:15.660 [2024-11-20 13:49:22.510856] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:29:15.660 [2024-11-20 13:49:22.510991] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:29:15.660 [2024-11-20 13:49:22.511216] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008580 00:29:15.660 [2024-11-20 13:49:22.511296] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:29:15.660 [2024-11-20 13:49:22.511549] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000068a0 00:29:15.660 [2024-11-20 13:49:22.511739] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008580 00:29:15.660 [2024-11-20 13:49:22.511816] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008580 00:29:15.660 [2024-11-20 13:49:22.512020] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:15.660 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:15.660 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:29:15.660 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:15.660 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:15.660 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:29:15.660 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:15.660 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:15.660 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:15.660 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:15.660 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:15.660 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:15.660 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:15.660 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:15.660 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:15.660 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:15.660 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:15.660 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:15.660 "name": "raid_bdev1", 00:29:15.660 "uuid": "c54e123e-4c58-4fd4-abfa-265082d71c53", 00:29:15.660 "strip_size_kb": 64, 00:29:15.660 "state": "online", 00:29:15.660 "raid_level": "raid0", 00:29:15.660 "superblock": true, 00:29:15.660 "num_base_bdevs": 4, 00:29:15.660 "num_base_bdevs_discovered": 4, 00:29:15.660 "num_base_bdevs_operational": 4, 00:29:15.660 "base_bdevs_list": [ 00:29:15.660 { 00:29:15.660 "name": "BaseBdev1", 00:29:15.660 "uuid": "18a7a6df-cbdb-5f65-a144-be526953a4dc", 00:29:15.660 "is_configured": true, 00:29:15.660 "data_offset": 2048, 00:29:15.660 "data_size": 63488 00:29:15.660 }, 00:29:15.660 { 00:29:15.660 "name": "BaseBdev2", 00:29:15.660 "uuid": "fb31dd80-15d2-5a9a-8e1e-1e74bebf2a0a", 00:29:15.660 "is_configured": true, 00:29:15.660 "data_offset": 2048, 00:29:15.660 "data_size": 63488 00:29:15.660 }, 00:29:15.660 { 00:29:15.660 "name": "BaseBdev3", 00:29:15.660 "uuid": "31452212-30c1-586a-8e7c-0efbec3ea747", 00:29:15.660 "is_configured": true, 00:29:15.660 "data_offset": 2048, 00:29:15.660 "data_size": 63488 00:29:15.660 }, 00:29:15.660 { 00:29:15.660 "name": "BaseBdev4", 00:29:15.660 "uuid": "59caa9ca-9a32-590f-9346-ff7fbf3daee7", 00:29:15.660 "is_configured": true, 00:29:15.660 "data_offset": 2048, 00:29:15.660 "data_size": 63488 00:29:15.660 } 00:29:15.660 ] 00:29:15.660 }' 00:29:15.660 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:15.660 13:49:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:15.919 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:29:15.919 13:49:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:29:15.919 [2024-11-20 13:49:22.925833] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006a40 00:29:16.934 13:49:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:29:16.934 13:49:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:16.934 13:49:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:16.934 13:49:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:16.934 13:49:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:29:16.934 13:49:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:29:16.934 13:49:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:29:16.934 13:49:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:29:16.934 13:49:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:16.934 13:49:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:16.934 13:49:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:29:16.934 13:49:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:16.934 13:49:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:16.934 13:49:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:16.934 13:49:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:16.934 13:49:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:16.934 13:49:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:16.934 13:49:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:16.934 13:49:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:16.934 13:49:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:16.934 13:49:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:16.934 13:49:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:16.934 13:49:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:16.934 "name": "raid_bdev1", 00:29:16.934 "uuid": "c54e123e-4c58-4fd4-abfa-265082d71c53", 00:29:16.934 "strip_size_kb": 64, 00:29:16.934 "state": "online", 00:29:16.934 "raid_level": "raid0", 00:29:16.934 "superblock": true, 00:29:16.934 "num_base_bdevs": 4, 00:29:16.934 "num_base_bdevs_discovered": 4, 00:29:16.934 "num_base_bdevs_operational": 4, 00:29:16.934 "base_bdevs_list": [ 00:29:16.934 { 00:29:16.934 "name": "BaseBdev1", 00:29:16.934 "uuid": "18a7a6df-cbdb-5f65-a144-be526953a4dc", 00:29:16.934 "is_configured": true, 00:29:16.934 "data_offset": 2048, 00:29:16.934 "data_size": 63488 00:29:16.934 }, 00:29:16.934 { 00:29:16.934 "name": "BaseBdev2", 00:29:16.934 "uuid": "fb31dd80-15d2-5a9a-8e1e-1e74bebf2a0a", 00:29:16.934 "is_configured": true, 00:29:16.934 "data_offset": 2048, 00:29:16.934 "data_size": 63488 00:29:16.934 }, 00:29:16.934 { 00:29:16.934 "name": "BaseBdev3", 00:29:16.934 "uuid": "31452212-30c1-586a-8e7c-0efbec3ea747", 00:29:16.934 "is_configured": true, 00:29:16.934 "data_offset": 2048, 00:29:16.934 "data_size": 63488 00:29:16.934 }, 00:29:16.934 { 00:29:16.934 "name": "BaseBdev4", 00:29:16.934 "uuid": "59caa9ca-9a32-590f-9346-ff7fbf3daee7", 00:29:16.934 "is_configured": true, 00:29:16.934 "data_offset": 2048, 00:29:16.934 "data_size": 63488 00:29:16.934 } 00:29:16.934 ] 00:29:16.934 }' 00:29:16.935 13:49:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:16.935 13:49:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:17.193 13:49:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:29:17.193 13:49:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:17.193 13:49:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:17.193 [2024-11-20 13:49:24.167182] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:29:17.193 [2024-11-20 13:49:24.167216] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:29:17.193 [2024-11-20 13:49:24.169746] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:17.193 [2024-11-20 13:49:24.169945] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:17.193 [2024-11-20 13:49:24.170004] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:29:17.193 [2024-11-20 13:49:24.170016] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state offline 00:29:17.193 { 00:29:17.193 "results": [ 00:29:17.193 { 00:29:17.193 "job": "raid_bdev1", 00:29:17.193 "core_mask": "0x1", 00:29:17.193 "workload": "randrw", 00:29:17.193 "percentage": 50, 00:29:17.193 "status": "finished", 00:29:17.193 "queue_depth": 1, 00:29:17.193 "io_size": 131072, 00:29:17.193 "runtime": 1.239552, 00:29:17.193 "iops": 16096.13795952086, 00:29:17.193 "mibps": 2012.0172449401075, 00:29:17.193 "io_failed": 1, 00:29:17.193 "io_timeout": 0, 00:29:17.193 "avg_latency_us": 85.53373535500734, 00:29:17.193 "min_latency_us": 27.56923076923077, 00:29:17.193 "max_latency_us": 1373.7353846153846 00:29:17.193 } 00:29:17.193 ], 00:29:17.193 "core_count": 1 00:29:17.193 } 00:29:17.193 13:49:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:17.193 13:49:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 69138 00:29:17.193 13:49:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 69138 ']' 00:29:17.193 13:49:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 69138 00:29:17.193 13:49:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:29:17.193 13:49:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:29:17.193 13:49:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69138 00:29:17.193 killing process with pid 69138 00:29:17.193 13:49:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:29:17.193 13:49:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:29:17.193 13:49:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69138' 00:29:17.193 13:49:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 69138 00:29:17.193 13:49:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 69138 00:29:17.193 [2024-11-20 13:49:24.200167] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:29:17.451 [2024-11-20 13:49:24.370712] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:29:18.016 13:49:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:29:18.016 13:49:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.UZKJbgzqP5 00:29:18.016 13:49:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:29:18.016 13:49:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.81 00:29:18.016 13:49:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:29:18.016 13:49:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:29:18.016 13:49:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:29:18.016 13:49:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.81 != \0\.\0\0 ]] 00:29:18.016 00:29:18.016 real 0m3.639s 00:29:18.016 user 0m4.285s 00:29:18.016 sys 0m0.440s 00:29:18.016 13:49:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:29:18.016 ************************************ 00:29:18.016 END TEST raid_read_error_test 00:29:18.016 ************************************ 00:29:18.016 13:49:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:18.275 13:49:25 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 4 write 00:29:18.275 13:49:25 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:29:18.275 13:49:25 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:29:18.275 13:49:25 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:29:18.275 ************************************ 00:29:18.275 START TEST raid_write_error_test 00:29:18.275 ************************************ 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 4 write 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.IfX3v0BV5u 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=69273 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 69273 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 69273 ']' 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:18.275 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:29:18.275 13:49:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:18.275 [2024-11-20 13:49:25.163924] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:29:18.275 [2024-11-20 13:49:25.164085] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69273 ] 00:29:18.275 [2024-11-20 13:49:25.325793] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:18.533 [2024-11-20 13:49:25.465288] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:29:18.792 [2024-11-20 13:49:25.637717] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:18.792 [2024-11-20 13:49:25.637776] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:19.051 BaseBdev1_malloc 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:19.051 true 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:19.051 [2024-11-20 13:49:26.060054] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:29:19.051 [2024-11-20 13:49:26.060250] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:19.051 [2024-11-20 13:49:26.060297] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:29:19.051 [2024-11-20 13:49:26.060388] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:19.051 [2024-11-20 13:49:26.062719] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:19.051 [2024-11-20 13:49:26.062848] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:29:19.051 BaseBdev1 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:19.051 BaseBdev2_malloc 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:19.051 true 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:19.051 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:19.311 [2024-11-20 13:49:26.110836] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:29:19.311 [2024-11-20 13:49:26.111004] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:19.311 [2024-11-20 13:49:26.111045] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:29:19.311 [2024-11-20 13:49:26.111548] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:19.311 [2024-11-20 13:49:26.117627] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:19.311 [2024-11-20 13:49:26.117722] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:29:19.311 BaseBdev2 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:19.311 BaseBdev3_malloc 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:19.311 true 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:19.311 [2024-11-20 13:49:26.178158] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:29:19.311 [2024-11-20 13:49:26.178309] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:19.311 [2024-11-20 13:49:26.178350] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:29:19.311 [2024-11-20 13:49:26.178409] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:19.311 [2024-11-20 13:49:26.180667] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:19.311 [2024-11-20 13:49:26.180772] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:29:19.311 BaseBdev3 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:19.311 BaseBdev4_malloc 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:19.311 true 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:19.311 [2024-11-20 13:49:26.224083] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:29:19.311 [2024-11-20 13:49:26.224132] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:19.311 [2024-11-20 13:49:26.224150] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:29:19.311 [2024-11-20 13:49:26.224161] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:19.311 [2024-11-20 13:49:26.226353] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:19.311 [2024-11-20 13:49:26.226388] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:29:19.311 BaseBdev4 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:19.311 [2024-11-20 13:49:26.232153] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:29:19.311 [2024-11-20 13:49:26.234113] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:29:19.311 [2024-11-20 13:49:26.234190] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:29:19.311 [2024-11-20 13:49:26.234256] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:29:19.311 [2024-11-20 13:49:26.234473] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008580 00:29:19.311 [2024-11-20 13:49:26.234487] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:29:19.311 [2024-11-20 13:49:26.234732] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000068a0 00:29:19.311 [2024-11-20 13:49:26.234882] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008580 00:29:19.311 [2024-11-20 13:49:26.234892] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008580 00:29:19.311 [2024-11-20 13:49:26.235046] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:19.311 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:19.312 "name": "raid_bdev1", 00:29:19.312 "uuid": "070aff61-8f34-458a-bad1-88d158570a75", 00:29:19.312 "strip_size_kb": 64, 00:29:19.312 "state": "online", 00:29:19.312 "raid_level": "raid0", 00:29:19.312 "superblock": true, 00:29:19.312 "num_base_bdevs": 4, 00:29:19.312 "num_base_bdevs_discovered": 4, 00:29:19.312 "num_base_bdevs_operational": 4, 00:29:19.312 "base_bdevs_list": [ 00:29:19.312 { 00:29:19.312 "name": "BaseBdev1", 00:29:19.312 "uuid": "39fa53b1-ab66-5b59-bc2c-dbeec9a529fe", 00:29:19.312 "is_configured": true, 00:29:19.312 "data_offset": 2048, 00:29:19.312 "data_size": 63488 00:29:19.312 }, 00:29:19.312 { 00:29:19.312 "name": "BaseBdev2", 00:29:19.312 "uuid": "bb6baa46-75c5-5be4-bb15-49c7113699d0", 00:29:19.312 "is_configured": true, 00:29:19.312 "data_offset": 2048, 00:29:19.312 "data_size": 63488 00:29:19.312 }, 00:29:19.312 { 00:29:19.312 "name": "BaseBdev3", 00:29:19.312 "uuid": "4ccb27ef-5dd9-5603-947b-80960c0e37f4", 00:29:19.312 "is_configured": true, 00:29:19.312 "data_offset": 2048, 00:29:19.312 "data_size": 63488 00:29:19.312 }, 00:29:19.312 { 00:29:19.312 "name": "BaseBdev4", 00:29:19.312 "uuid": "7fba2736-f42c-5602-a7e2-8495a2426984", 00:29:19.312 "is_configured": true, 00:29:19.312 "data_offset": 2048, 00:29:19.312 "data_size": 63488 00:29:19.312 } 00:29:19.312 ] 00:29:19.312 }' 00:29:19.312 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:19.312 13:49:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:19.570 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:29:19.570 13:49:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:29:19.828 [2024-11-20 13:49:26.645293] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006a40 00:29:20.870 13:49:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:29:20.870 13:49:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:20.870 13:49:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:20.870 13:49:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:20.870 13:49:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:29:20.870 13:49:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:29:20.870 13:49:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:29:20.870 13:49:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:29:20.870 13:49:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:20.870 13:49:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:20.870 13:49:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:29:20.870 13:49:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:20.870 13:49:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:20.870 13:49:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:20.870 13:49:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:20.870 13:49:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:20.870 13:49:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:20.870 13:49:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:20.870 13:49:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:20.870 13:49:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:20.870 13:49:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:20.870 13:49:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:20.870 13:49:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:20.870 "name": "raid_bdev1", 00:29:20.870 "uuid": "070aff61-8f34-458a-bad1-88d158570a75", 00:29:20.870 "strip_size_kb": 64, 00:29:20.870 "state": "online", 00:29:20.870 "raid_level": "raid0", 00:29:20.870 "superblock": true, 00:29:20.871 "num_base_bdevs": 4, 00:29:20.871 "num_base_bdevs_discovered": 4, 00:29:20.871 "num_base_bdevs_operational": 4, 00:29:20.871 "base_bdevs_list": [ 00:29:20.871 { 00:29:20.871 "name": "BaseBdev1", 00:29:20.871 "uuid": "39fa53b1-ab66-5b59-bc2c-dbeec9a529fe", 00:29:20.871 "is_configured": true, 00:29:20.871 "data_offset": 2048, 00:29:20.871 "data_size": 63488 00:29:20.871 }, 00:29:20.871 { 00:29:20.871 "name": "BaseBdev2", 00:29:20.871 "uuid": "bb6baa46-75c5-5be4-bb15-49c7113699d0", 00:29:20.871 "is_configured": true, 00:29:20.871 "data_offset": 2048, 00:29:20.871 "data_size": 63488 00:29:20.871 }, 00:29:20.871 { 00:29:20.871 "name": "BaseBdev3", 00:29:20.871 "uuid": "4ccb27ef-5dd9-5603-947b-80960c0e37f4", 00:29:20.871 "is_configured": true, 00:29:20.871 "data_offset": 2048, 00:29:20.871 "data_size": 63488 00:29:20.871 }, 00:29:20.871 { 00:29:20.871 "name": "BaseBdev4", 00:29:20.871 "uuid": "7fba2736-f42c-5602-a7e2-8495a2426984", 00:29:20.871 "is_configured": true, 00:29:20.871 "data_offset": 2048, 00:29:20.871 "data_size": 63488 00:29:20.871 } 00:29:20.871 ] 00:29:20.871 }' 00:29:20.871 13:49:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:20.871 13:49:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:20.871 13:49:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:29:20.871 13:49:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:20.871 13:49:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:20.871 [2024-11-20 13:49:27.891787] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:29:20.871 [2024-11-20 13:49:27.891824] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:29:20.871 [2024-11-20 13:49:27.895119] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:20.871 [2024-11-20 13:49:27.895263] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:20.871 [2024-11-20 13:49:27.895367] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:29:20.871 [2024-11-20 13:49:27.895439] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state offline 00:29:20.871 { 00:29:20.871 "results": [ 00:29:20.871 { 00:29:20.871 "job": "raid_bdev1", 00:29:20.871 "core_mask": "0x1", 00:29:20.871 "workload": "randrw", 00:29:20.871 "percentage": 50, 00:29:20.871 "status": "finished", 00:29:20.871 "queue_depth": 1, 00:29:20.871 "io_size": 131072, 00:29:20.871 "runtime": 1.244395, 00:29:20.871 "iops": 13637.952579365876, 00:29:20.871 "mibps": 1704.7440724207345, 00:29:20.871 "io_failed": 1, 00:29:20.871 "io_timeout": 0, 00:29:20.871 "avg_latency_us": 100.45281459054733, 00:29:20.871 "min_latency_us": 34.067692307692305, 00:29:20.871 "max_latency_us": 1714.0184615384615 00:29:20.871 } 00:29:20.871 ], 00:29:20.871 "core_count": 1 00:29:20.871 } 00:29:20.871 13:49:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:20.871 13:49:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 69273 00:29:20.871 13:49:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 69273 ']' 00:29:20.871 13:49:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 69273 00:29:20.871 13:49:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:29:20.871 13:49:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:29:20.871 13:49:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69273 00:29:21.129 killing process with pid 69273 00:29:21.129 13:49:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:29:21.129 13:49:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:29:21.129 13:49:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69273' 00:29:21.129 13:49:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 69273 00:29:21.129 [2024-11-20 13:49:27.928507] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:29:21.129 13:49:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 69273 00:29:21.129 [2024-11-20 13:49:28.144811] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:29:22.064 13:49:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:29:22.064 13:49:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:29:22.064 13:49:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.IfX3v0BV5u 00:29:22.064 ************************************ 00:29:22.064 END TEST raid_write_error_test 00:29:22.064 ************************************ 00:29:22.064 13:49:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.80 00:29:22.064 13:49:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:29:22.064 13:49:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:29:22.064 13:49:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:29:22.064 13:49:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.80 != \0\.\0\0 ]] 00:29:22.064 00:29:22.064 real 0m3.887s 00:29:22.064 user 0m4.494s 00:29:22.064 sys 0m0.498s 00:29:22.064 13:49:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:29:22.064 13:49:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:22.064 13:49:28 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:29:22.064 13:49:28 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 4 false 00:29:22.064 13:49:28 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:29:22.064 13:49:28 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:29:22.064 13:49:28 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:29:22.064 ************************************ 00:29:22.064 START TEST raid_state_function_test 00:29:22.064 ************************************ 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 4 false 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:29:22.064 Process raid pid: 69411 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=69411 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 69411' 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 69411 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 69411 ']' 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:22.064 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:22.064 13:49:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:29:22.065 13:49:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:22.065 13:49:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:29:22.065 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:29:22.065 13:49:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:22.065 [2024-11-20 13:49:29.081321] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:29:22.065 [2024-11-20 13:49:29.081591] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:29:22.327 [2024-11-20 13:49:29.244794] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:22.327 [2024-11-20 13:49:29.361489] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:29:22.584 [2024-11-20 13:49:29.510025] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:22.584 [2024-11-20 13:49:29.510251] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:23.150 13:49:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:29:23.150 13:49:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:29:23.150 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:29:23.150 13:49:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:23.150 13:49:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:23.150 [2024-11-20 13:49:29.946591] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:29:23.150 [2024-11-20 13:49:29.946654] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:29:23.150 [2024-11-20 13:49:29.946666] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:29:23.150 [2024-11-20 13:49:29.946676] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:29:23.150 [2024-11-20 13:49:29.946683] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:29:23.150 [2024-11-20 13:49:29.946692] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:29:23.150 [2024-11-20 13:49:29.946698] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:29:23.150 [2024-11-20 13:49:29.946707] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:29:23.150 13:49:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:23.150 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:29:23.150 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:23.150 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:23.150 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:23.150 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:23.150 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:23.150 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:23.150 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:23.150 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:23.150 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:23.150 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:23.150 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:23.150 13:49:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:23.150 13:49:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:23.150 13:49:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:23.150 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:23.150 "name": "Existed_Raid", 00:29:23.150 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:23.150 "strip_size_kb": 64, 00:29:23.150 "state": "configuring", 00:29:23.150 "raid_level": "concat", 00:29:23.150 "superblock": false, 00:29:23.150 "num_base_bdevs": 4, 00:29:23.150 "num_base_bdevs_discovered": 0, 00:29:23.150 "num_base_bdevs_operational": 4, 00:29:23.150 "base_bdevs_list": [ 00:29:23.150 { 00:29:23.150 "name": "BaseBdev1", 00:29:23.150 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:23.150 "is_configured": false, 00:29:23.150 "data_offset": 0, 00:29:23.150 "data_size": 0 00:29:23.150 }, 00:29:23.151 { 00:29:23.151 "name": "BaseBdev2", 00:29:23.151 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:23.151 "is_configured": false, 00:29:23.151 "data_offset": 0, 00:29:23.151 "data_size": 0 00:29:23.151 }, 00:29:23.151 { 00:29:23.151 "name": "BaseBdev3", 00:29:23.151 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:23.151 "is_configured": false, 00:29:23.151 "data_offset": 0, 00:29:23.151 "data_size": 0 00:29:23.151 }, 00:29:23.151 { 00:29:23.151 "name": "BaseBdev4", 00:29:23.151 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:23.151 "is_configured": false, 00:29:23.151 "data_offset": 0, 00:29:23.151 "data_size": 0 00:29:23.151 } 00:29:23.151 ] 00:29:23.151 }' 00:29:23.151 13:49:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:23.151 13:49:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:23.409 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:29:23.409 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:23.409 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:23.409 [2024-11-20 13:49:30.262871] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:29:23.409 [2024-11-20 13:49:30.262915] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:29:23.409 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:23.409 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:29:23.409 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:23.409 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:23.409 [2024-11-20 13:49:30.270609] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:29:23.409 [2024-11-20 13:49:30.270653] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:29:23.409 [2024-11-20 13:49:30.270663] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:29:23.409 [2024-11-20 13:49:30.270675] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:29:23.409 [2024-11-20 13:49:30.270682] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:29:23.409 [2024-11-20 13:49:30.270692] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:29:23.409 [2024-11-20 13:49:30.270699] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:29:23.409 [2024-11-20 13:49:30.270709] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:29:23.409 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:23.409 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:29:23.409 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:23.409 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:23.409 [2024-11-20 13:49:30.305483] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:29:23.409 BaseBdev1 00:29:23.409 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:23.410 [ 00:29:23.410 { 00:29:23.410 "name": "BaseBdev1", 00:29:23.410 "aliases": [ 00:29:23.410 "607ff823-992b-4ed7-b218-f8d566153a71" 00:29:23.410 ], 00:29:23.410 "product_name": "Malloc disk", 00:29:23.410 "block_size": 512, 00:29:23.410 "num_blocks": 65536, 00:29:23.410 "uuid": "607ff823-992b-4ed7-b218-f8d566153a71", 00:29:23.410 "assigned_rate_limits": { 00:29:23.410 "rw_ios_per_sec": 0, 00:29:23.410 "rw_mbytes_per_sec": 0, 00:29:23.410 "r_mbytes_per_sec": 0, 00:29:23.410 "w_mbytes_per_sec": 0 00:29:23.410 }, 00:29:23.410 "claimed": true, 00:29:23.410 "claim_type": "exclusive_write", 00:29:23.410 "zoned": false, 00:29:23.410 "supported_io_types": { 00:29:23.410 "read": true, 00:29:23.410 "write": true, 00:29:23.410 "unmap": true, 00:29:23.410 "flush": true, 00:29:23.410 "reset": true, 00:29:23.410 "nvme_admin": false, 00:29:23.410 "nvme_io": false, 00:29:23.410 "nvme_io_md": false, 00:29:23.410 "write_zeroes": true, 00:29:23.410 "zcopy": true, 00:29:23.410 "get_zone_info": false, 00:29:23.410 "zone_management": false, 00:29:23.410 "zone_append": false, 00:29:23.410 "compare": false, 00:29:23.410 "compare_and_write": false, 00:29:23.410 "abort": true, 00:29:23.410 "seek_hole": false, 00:29:23.410 "seek_data": false, 00:29:23.410 "copy": true, 00:29:23.410 "nvme_iov_md": false 00:29:23.410 }, 00:29:23.410 "memory_domains": [ 00:29:23.410 { 00:29:23.410 "dma_device_id": "system", 00:29:23.410 "dma_device_type": 1 00:29:23.410 }, 00:29:23.410 { 00:29:23.410 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:23.410 "dma_device_type": 2 00:29:23.410 } 00:29:23.410 ], 00:29:23.410 "driver_specific": {} 00:29:23.410 } 00:29:23.410 ] 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:23.410 "name": "Existed_Raid", 00:29:23.410 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:23.410 "strip_size_kb": 64, 00:29:23.410 "state": "configuring", 00:29:23.410 "raid_level": "concat", 00:29:23.410 "superblock": false, 00:29:23.410 "num_base_bdevs": 4, 00:29:23.410 "num_base_bdevs_discovered": 1, 00:29:23.410 "num_base_bdevs_operational": 4, 00:29:23.410 "base_bdevs_list": [ 00:29:23.410 { 00:29:23.410 "name": "BaseBdev1", 00:29:23.410 "uuid": "607ff823-992b-4ed7-b218-f8d566153a71", 00:29:23.410 "is_configured": true, 00:29:23.410 "data_offset": 0, 00:29:23.410 "data_size": 65536 00:29:23.410 }, 00:29:23.410 { 00:29:23.410 "name": "BaseBdev2", 00:29:23.410 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:23.410 "is_configured": false, 00:29:23.410 "data_offset": 0, 00:29:23.410 "data_size": 0 00:29:23.410 }, 00:29:23.410 { 00:29:23.410 "name": "BaseBdev3", 00:29:23.410 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:23.410 "is_configured": false, 00:29:23.410 "data_offset": 0, 00:29:23.410 "data_size": 0 00:29:23.410 }, 00:29:23.410 { 00:29:23.410 "name": "BaseBdev4", 00:29:23.410 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:23.410 "is_configured": false, 00:29:23.410 "data_offset": 0, 00:29:23.410 "data_size": 0 00:29:23.410 } 00:29:23.410 ] 00:29:23.410 }' 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:23.410 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:23.669 [2024-11-20 13:49:30.645629] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:29:23.669 [2024-11-20 13:49:30.645687] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:23.669 [2024-11-20 13:49:30.653679] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:29:23.669 [2024-11-20 13:49:30.655737] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:29:23.669 [2024-11-20 13:49:30.655860] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:29:23.669 [2024-11-20 13:49:30.655946] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:29:23.669 [2024-11-20 13:49:30.655976] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:29:23.669 [2024-11-20 13:49:30.656041] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:29:23.669 [2024-11-20 13:49:30.656068] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:23.669 "name": "Existed_Raid", 00:29:23.669 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:23.669 "strip_size_kb": 64, 00:29:23.669 "state": "configuring", 00:29:23.669 "raid_level": "concat", 00:29:23.669 "superblock": false, 00:29:23.669 "num_base_bdevs": 4, 00:29:23.669 "num_base_bdevs_discovered": 1, 00:29:23.669 "num_base_bdevs_operational": 4, 00:29:23.669 "base_bdevs_list": [ 00:29:23.669 { 00:29:23.669 "name": "BaseBdev1", 00:29:23.669 "uuid": "607ff823-992b-4ed7-b218-f8d566153a71", 00:29:23.669 "is_configured": true, 00:29:23.669 "data_offset": 0, 00:29:23.669 "data_size": 65536 00:29:23.669 }, 00:29:23.669 { 00:29:23.669 "name": "BaseBdev2", 00:29:23.669 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:23.669 "is_configured": false, 00:29:23.669 "data_offset": 0, 00:29:23.669 "data_size": 0 00:29:23.669 }, 00:29:23.669 { 00:29:23.669 "name": "BaseBdev3", 00:29:23.669 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:23.669 "is_configured": false, 00:29:23.669 "data_offset": 0, 00:29:23.669 "data_size": 0 00:29:23.669 }, 00:29:23.669 { 00:29:23.669 "name": "BaseBdev4", 00:29:23.669 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:23.669 "is_configured": false, 00:29:23.669 "data_offset": 0, 00:29:23.669 "data_size": 0 00:29:23.669 } 00:29:23.669 ] 00:29:23.669 }' 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:23.669 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:23.940 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:29:23.940 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:23.940 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:23.940 [2024-11-20 13:49:30.978533] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:29:23.940 BaseBdev2 00:29:23.940 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:23.941 13:49:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:29:23.941 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:29:23.941 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:23.941 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:29:23.941 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:23.941 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:23.941 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:23.941 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:23.941 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:23.941 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:23.941 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:29:23.941 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:23.941 13:49:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.198 [ 00:29:24.198 { 00:29:24.198 "name": "BaseBdev2", 00:29:24.198 "aliases": [ 00:29:24.198 "1d9ed5a9-d3b5-4928-9f79-f6ecf1fb50cd" 00:29:24.198 ], 00:29:24.198 "product_name": "Malloc disk", 00:29:24.198 "block_size": 512, 00:29:24.198 "num_blocks": 65536, 00:29:24.198 "uuid": "1d9ed5a9-d3b5-4928-9f79-f6ecf1fb50cd", 00:29:24.198 "assigned_rate_limits": { 00:29:24.198 "rw_ios_per_sec": 0, 00:29:24.198 "rw_mbytes_per_sec": 0, 00:29:24.198 "r_mbytes_per_sec": 0, 00:29:24.198 "w_mbytes_per_sec": 0 00:29:24.198 }, 00:29:24.198 "claimed": true, 00:29:24.198 "claim_type": "exclusive_write", 00:29:24.198 "zoned": false, 00:29:24.198 "supported_io_types": { 00:29:24.198 "read": true, 00:29:24.198 "write": true, 00:29:24.198 "unmap": true, 00:29:24.198 "flush": true, 00:29:24.198 "reset": true, 00:29:24.198 "nvme_admin": false, 00:29:24.198 "nvme_io": false, 00:29:24.198 "nvme_io_md": false, 00:29:24.198 "write_zeroes": true, 00:29:24.198 "zcopy": true, 00:29:24.198 "get_zone_info": false, 00:29:24.198 "zone_management": false, 00:29:24.198 "zone_append": false, 00:29:24.198 "compare": false, 00:29:24.198 "compare_and_write": false, 00:29:24.198 "abort": true, 00:29:24.198 "seek_hole": false, 00:29:24.198 "seek_data": false, 00:29:24.198 "copy": true, 00:29:24.198 "nvme_iov_md": false 00:29:24.198 }, 00:29:24.198 "memory_domains": [ 00:29:24.198 { 00:29:24.198 "dma_device_id": "system", 00:29:24.198 "dma_device_type": 1 00:29:24.198 }, 00:29:24.198 { 00:29:24.198 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:24.198 "dma_device_type": 2 00:29:24.198 } 00:29:24.198 ], 00:29:24.198 "driver_specific": {} 00:29:24.198 } 00:29:24.198 ] 00:29:24.198 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:24.198 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:29:24.198 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:29:24.198 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:29:24.198 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:29:24.198 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:24.198 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:24.198 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:24.198 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:24.198 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:24.199 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:24.199 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:24.199 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:24.199 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:24.199 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:24.199 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:24.199 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:24.199 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.199 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:24.199 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:24.199 "name": "Existed_Raid", 00:29:24.199 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:24.199 "strip_size_kb": 64, 00:29:24.199 "state": "configuring", 00:29:24.199 "raid_level": "concat", 00:29:24.199 "superblock": false, 00:29:24.199 "num_base_bdevs": 4, 00:29:24.199 "num_base_bdevs_discovered": 2, 00:29:24.199 "num_base_bdevs_operational": 4, 00:29:24.199 "base_bdevs_list": [ 00:29:24.199 { 00:29:24.199 "name": "BaseBdev1", 00:29:24.199 "uuid": "607ff823-992b-4ed7-b218-f8d566153a71", 00:29:24.199 "is_configured": true, 00:29:24.199 "data_offset": 0, 00:29:24.199 "data_size": 65536 00:29:24.199 }, 00:29:24.199 { 00:29:24.199 "name": "BaseBdev2", 00:29:24.199 "uuid": "1d9ed5a9-d3b5-4928-9f79-f6ecf1fb50cd", 00:29:24.199 "is_configured": true, 00:29:24.199 "data_offset": 0, 00:29:24.199 "data_size": 65536 00:29:24.199 }, 00:29:24.199 { 00:29:24.199 "name": "BaseBdev3", 00:29:24.199 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:24.199 "is_configured": false, 00:29:24.199 "data_offset": 0, 00:29:24.199 "data_size": 0 00:29:24.199 }, 00:29:24.199 { 00:29:24.199 "name": "BaseBdev4", 00:29:24.199 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:24.199 "is_configured": false, 00:29:24.199 "data_offset": 0, 00:29:24.199 "data_size": 0 00:29:24.199 } 00:29:24.199 ] 00:29:24.199 }' 00:29:24.199 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:24.199 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.457 [2024-11-20 13:49:31.342910] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:29:24.457 BaseBdev3 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.457 [ 00:29:24.457 { 00:29:24.457 "name": "BaseBdev3", 00:29:24.457 "aliases": [ 00:29:24.457 "41fe2c4a-b3c6-4498-b6c1-7c272b0f3bbc" 00:29:24.457 ], 00:29:24.457 "product_name": "Malloc disk", 00:29:24.457 "block_size": 512, 00:29:24.457 "num_blocks": 65536, 00:29:24.457 "uuid": "41fe2c4a-b3c6-4498-b6c1-7c272b0f3bbc", 00:29:24.457 "assigned_rate_limits": { 00:29:24.457 "rw_ios_per_sec": 0, 00:29:24.457 "rw_mbytes_per_sec": 0, 00:29:24.457 "r_mbytes_per_sec": 0, 00:29:24.457 "w_mbytes_per_sec": 0 00:29:24.457 }, 00:29:24.457 "claimed": true, 00:29:24.457 "claim_type": "exclusive_write", 00:29:24.457 "zoned": false, 00:29:24.457 "supported_io_types": { 00:29:24.457 "read": true, 00:29:24.457 "write": true, 00:29:24.457 "unmap": true, 00:29:24.457 "flush": true, 00:29:24.457 "reset": true, 00:29:24.457 "nvme_admin": false, 00:29:24.457 "nvme_io": false, 00:29:24.457 "nvme_io_md": false, 00:29:24.457 "write_zeroes": true, 00:29:24.457 "zcopy": true, 00:29:24.457 "get_zone_info": false, 00:29:24.457 "zone_management": false, 00:29:24.457 "zone_append": false, 00:29:24.457 "compare": false, 00:29:24.457 "compare_and_write": false, 00:29:24.457 "abort": true, 00:29:24.457 "seek_hole": false, 00:29:24.457 "seek_data": false, 00:29:24.457 "copy": true, 00:29:24.457 "nvme_iov_md": false 00:29:24.457 }, 00:29:24.457 "memory_domains": [ 00:29:24.457 { 00:29:24.457 "dma_device_id": "system", 00:29:24.457 "dma_device_type": 1 00:29:24.457 }, 00:29:24.457 { 00:29:24.457 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:24.457 "dma_device_type": 2 00:29:24.457 } 00:29:24.457 ], 00:29:24.457 "driver_specific": {} 00:29:24.457 } 00:29:24.457 ] 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:24.457 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:24.458 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:24.458 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:24.458 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:24.458 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:24.458 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.458 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:24.458 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:24.458 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:24.458 "name": "Existed_Raid", 00:29:24.458 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:24.458 "strip_size_kb": 64, 00:29:24.458 "state": "configuring", 00:29:24.458 "raid_level": "concat", 00:29:24.458 "superblock": false, 00:29:24.458 "num_base_bdevs": 4, 00:29:24.458 "num_base_bdevs_discovered": 3, 00:29:24.458 "num_base_bdevs_operational": 4, 00:29:24.458 "base_bdevs_list": [ 00:29:24.458 { 00:29:24.458 "name": "BaseBdev1", 00:29:24.458 "uuid": "607ff823-992b-4ed7-b218-f8d566153a71", 00:29:24.458 "is_configured": true, 00:29:24.458 "data_offset": 0, 00:29:24.458 "data_size": 65536 00:29:24.458 }, 00:29:24.458 { 00:29:24.458 "name": "BaseBdev2", 00:29:24.458 "uuid": "1d9ed5a9-d3b5-4928-9f79-f6ecf1fb50cd", 00:29:24.458 "is_configured": true, 00:29:24.458 "data_offset": 0, 00:29:24.458 "data_size": 65536 00:29:24.458 }, 00:29:24.458 { 00:29:24.458 "name": "BaseBdev3", 00:29:24.458 "uuid": "41fe2c4a-b3c6-4498-b6c1-7c272b0f3bbc", 00:29:24.458 "is_configured": true, 00:29:24.458 "data_offset": 0, 00:29:24.458 "data_size": 65536 00:29:24.458 }, 00:29:24.458 { 00:29:24.458 "name": "BaseBdev4", 00:29:24.458 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:24.458 "is_configured": false, 00:29:24.458 "data_offset": 0, 00:29:24.458 "data_size": 0 00:29:24.458 } 00:29:24.458 ] 00:29:24.458 }' 00:29:24.458 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:24.458 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.716 [2024-11-20 13:49:31.715634] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:29:24.716 [2024-11-20 13:49:31.715691] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:29:24.716 [2024-11-20 13:49:31.715700] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:29:24.716 [2024-11-20 13:49:31.715972] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:29:24.716 [2024-11-20 13:49:31.716161] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:29:24.716 [2024-11-20 13:49:31.716173] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:29:24.716 [2024-11-20 13:49:31.716441] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:24.716 BaseBdev4 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.716 [ 00:29:24.716 { 00:29:24.716 "name": "BaseBdev4", 00:29:24.716 "aliases": [ 00:29:24.716 "069bb9e3-801f-41b8-8c73-5c2effb137f7" 00:29:24.716 ], 00:29:24.716 "product_name": "Malloc disk", 00:29:24.716 "block_size": 512, 00:29:24.716 "num_blocks": 65536, 00:29:24.716 "uuid": "069bb9e3-801f-41b8-8c73-5c2effb137f7", 00:29:24.716 "assigned_rate_limits": { 00:29:24.716 "rw_ios_per_sec": 0, 00:29:24.716 "rw_mbytes_per_sec": 0, 00:29:24.716 "r_mbytes_per_sec": 0, 00:29:24.716 "w_mbytes_per_sec": 0 00:29:24.716 }, 00:29:24.716 "claimed": true, 00:29:24.716 "claim_type": "exclusive_write", 00:29:24.716 "zoned": false, 00:29:24.716 "supported_io_types": { 00:29:24.716 "read": true, 00:29:24.716 "write": true, 00:29:24.716 "unmap": true, 00:29:24.716 "flush": true, 00:29:24.716 "reset": true, 00:29:24.716 "nvme_admin": false, 00:29:24.716 "nvme_io": false, 00:29:24.716 "nvme_io_md": false, 00:29:24.716 "write_zeroes": true, 00:29:24.716 "zcopy": true, 00:29:24.716 "get_zone_info": false, 00:29:24.716 "zone_management": false, 00:29:24.716 "zone_append": false, 00:29:24.716 "compare": false, 00:29:24.716 "compare_and_write": false, 00:29:24.716 "abort": true, 00:29:24.716 "seek_hole": false, 00:29:24.716 "seek_data": false, 00:29:24.716 "copy": true, 00:29:24.716 "nvme_iov_md": false 00:29:24.716 }, 00:29:24.716 "memory_domains": [ 00:29:24.716 { 00:29:24.716 "dma_device_id": "system", 00:29:24.716 "dma_device_type": 1 00:29:24.716 }, 00:29:24.716 { 00:29:24.716 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:24.716 "dma_device_type": 2 00:29:24.716 } 00:29:24.716 ], 00:29:24.716 "driver_specific": {} 00:29:24.716 } 00:29:24.716 ] 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:24.716 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:24.717 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:24.717 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:24.717 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:24.717 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.717 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:24.974 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:24.974 "name": "Existed_Raid", 00:29:24.974 "uuid": "2c2b1900-07bd-418e-8672-e6af1c48621f", 00:29:24.974 "strip_size_kb": 64, 00:29:24.974 "state": "online", 00:29:24.974 "raid_level": "concat", 00:29:24.974 "superblock": false, 00:29:24.974 "num_base_bdevs": 4, 00:29:24.974 "num_base_bdevs_discovered": 4, 00:29:24.974 "num_base_bdevs_operational": 4, 00:29:24.974 "base_bdevs_list": [ 00:29:24.975 { 00:29:24.975 "name": "BaseBdev1", 00:29:24.975 "uuid": "607ff823-992b-4ed7-b218-f8d566153a71", 00:29:24.975 "is_configured": true, 00:29:24.975 "data_offset": 0, 00:29:24.975 "data_size": 65536 00:29:24.975 }, 00:29:24.975 { 00:29:24.975 "name": "BaseBdev2", 00:29:24.975 "uuid": "1d9ed5a9-d3b5-4928-9f79-f6ecf1fb50cd", 00:29:24.975 "is_configured": true, 00:29:24.975 "data_offset": 0, 00:29:24.975 "data_size": 65536 00:29:24.975 }, 00:29:24.975 { 00:29:24.975 "name": "BaseBdev3", 00:29:24.975 "uuid": "41fe2c4a-b3c6-4498-b6c1-7c272b0f3bbc", 00:29:24.975 "is_configured": true, 00:29:24.975 "data_offset": 0, 00:29:24.975 "data_size": 65536 00:29:24.975 }, 00:29:24.975 { 00:29:24.975 "name": "BaseBdev4", 00:29:24.975 "uuid": "069bb9e3-801f-41b8-8c73-5c2effb137f7", 00:29:24.975 "is_configured": true, 00:29:24.975 "data_offset": 0, 00:29:24.975 "data_size": 65536 00:29:24.975 } 00:29:24.975 ] 00:29:24.975 }' 00:29:24.975 13:49:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:24.975 13:49:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:25.233 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:29:25.233 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:29:25.233 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:29:25.233 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:29:25.233 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:29:25.233 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:29:25.233 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:29:25.233 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:29:25.233 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:25.233 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:25.233 [2024-11-20 13:49:32.076183] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:25.233 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:25.233 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:25.233 "name": "Existed_Raid", 00:29:25.233 "aliases": [ 00:29:25.233 "2c2b1900-07bd-418e-8672-e6af1c48621f" 00:29:25.233 ], 00:29:25.233 "product_name": "Raid Volume", 00:29:25.233 "block_size": 512, 00:29:25.233 "num_blocks": 262144, 00:29:25.233 "uuid": "2c2b1900-07bd-418e-8672-e6af1c48621f", 00:29:25.233 "assigned_rate_limits": { 00:29:25.233 "rw_ios_per_sec": 0, 00:29:25.233 "rw_mbytes_per_sec": 0, 00:29:25.233 "r_mbytes_per_sec": 0, 00:29:25.233 "w_mbytes_per_sec": 0 00:29:25.233 }, 00:29:25.233 "claimed": false, 00:29:25.233 "zoned": false, 00:29:25.233 "supported_io_types": { 00:29:25.233 "read": true, 00:29:25.233 "write": true, 00:29:25.233 "unmap": true, 00:29:25.233 "flush": true, 00:29:25.233 "reset": true, 00:29:25.233 "nvme_admin": false, 00:29:25.233 "nvme_io": false, 00:29:25.233 "nvme_io_md": false, 00:29:25.233 "write_zeroes": true, 00:29:25.233 "zcopy": false, 00:29:25.233 "get_zone_info": false, 00:29:25.233 "zone_management": false, 00:29:25.233 "zone_append": false, 00:29:25.233 "compare": false, 00:29:25.234 "compare_and_write": false, 00:29:25.234 "abort": false, 00:29:25.234 "seek_hole": false, 00:29:25.234 "seek_data": false, 00:29:25.234 "copy": false, 00:29:25.234 "nvme_iov_md": false 00:29:25.234 }, 00:29:25.234 "memory_domains": [ 00:29:25.234 { 00:29:25.234 "dma_device_id": "system", 00:29:25.234 "dma_device_type": 1 00:29:25.234 }, 00:29:25.234 { 00:29:25.234 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:25.234 "dma_device_type": 2 00:29:25.234 }, 00:29:25.234 { 00:29:25.234 "dma_device_id": "system", 00:29:25.234 "dma_device_type": 1 00:29:25.234 }, 00:29:25.234 { 00:29:25.234 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:25.234 "dma_device_type": 2 00:29:25.234 }, 00:29:25.234 { 00:29:25.234 "dma_device_id": "system", 00:29:25.234 "dma_device_type": 1 00:29:25.234 }, 00:29:25.234 { 00:29:25.234 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:25.234 "dma_device_type": 2 00:29:25.234 }, 00:29:25.234 { 00:29:25.234 "dma_device_id": "system", 00:29:25.234 "dma_device_type": 1 00:29:25.234 }, 00:29:25.234 { 00:29:25.234 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:25.234 "dma_device_type": 2 00:29:25.234 } 00:29:25.234 ], 00:29:25.234 "driver_specific": { 00:29:25.234 "raid": { 00:29:25.234 "uuid": "2c2b1900-07bd-418e-8672-e6af1c48621f", 00:29:25.234 "strip_size_kb": 64, 00:29:25.234 "state": "online", 00:29:25.234 "raid_level": "concat", 00:29:25.234 "superblock": false, 00:29:25.234 "num_base_bdevs": 4, 00:29:25.234 "num_base_bdevs_discovered": 4, 00:29:25.234 "num_base_bdevs_operational": 4, 00:29:25.234 "base_bdevs_list": [ 00:29:25.234 { 00:29:25.234 "name": "BaseBdev1", 00:29:25.234 "uuid": "607ff823-992b-4ed7-b218-f8d566153a71", 00:29:25.234 "is_configured": true, 00:29:25.234 "data_offset": 0, 00:29:25.234 "data_size": 65536 00:29:25.234 }, 00:29:25.234 { 00:29:25.234 "name": "BaseBdev2", 00:29:25.234 "uuid": "1d9ed5a9-d3b5-4928-9f79-f6ecf1fb50cd", 00:29:25.234 "is_configured": true, 00:29:25.234 "data_offset": 0, 00:29:25.234 "data_size": 65536 00:29:25.234 }, 00:29:25.234 { 00:29:25.234 "name": "BaseBdev3", 00:29:25.234 "uuid": "41fe2c4a-b3c6-4498-b6c1-7c272b0f3bbc", 00:29:25.234 "is_configured": true, 00:29:25.234 "data_offset": 0, 00:29:25.234 "data_size": 65536 00:29:25.234 }, 00:29:25.234 { 00:29:25.234 "name": "BaseBdev4", 00:29:25.234 "uuid": "069bb9e3-801f-41b8-8c73-5c2effb137f7", 00:29:25.234 "is_configured": true, 00:29:25.234 "data_offset": 0, 00:29:25.234 "data_size": 65536 00:29:25.234 } 00:29:25.234 ] 00:29:25.234 } 00:29:25.234 } 00:29:25.234 }' 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:29:25.234 BaseBdev2 00:29:25.234 BaseBdev3 00:29:25.234 BaseBdev4' 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:25.234 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:25.493 [2024-11-20 13:49:32.299921] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:29:25.493 [2024-11-20 13:49:32.299960] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:29:25.493 [2024-11-20 13:49:32.300034] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 3 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:25.493 "name": "Existed_Raid", 00:29:25.493 "uuid": "2c2b1900-07bd-418e-8672-e6af1c48621f", 00:29:25.493 "strip_size_kb": 64, 00:29:25.493 "state": "offline", 00:29:25.493 "raid_level": "concat", 00:29:25.493 "superblock": false, 00:29:25.493 "num_base_bdevs": 4, 00:29:25.493 "num_base_bdevs_discovered": 3, 00:29:25.493 "num_base_bdevs_operational": 3, 00:29:25.493 "base_bdevs_list": [ 00:29:25.493 { 00:29:25.493 "name": null, 00:29:25.493 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:25.493 "is_configured": false, 00:29:25.493 "data_offset": 0, 00:29:25.493 "data_size": 65536 00:29:25.493 }, 00:29:25.493 { 00:29:25.493 "name": "BaseBdev2", 00:29:25.493 "uuid": "1d9ed5a9-d3b5-4928-9f79-f6ecf1fb50cd", 00:29:25.493 "is_configured": true, 00:29:25.493 "data_offset": 0, 00:29:25.493 "data_size": 65536 00:29:25.493 }, 00:29:25.493 { 00:29:25.493 "name": "BaseBdev3", 00:29:25.493 "uuid": "41fe2c4a-b3c6-4498-b6c1-7c272b0f3bbc", 00:29:25.493 "is_configured": true, 00:29:25.493 "data_offset": 0, 00:29:25.493 "data_size": 65536 00:29:25.493 }, 00:29:25.493 { 00:29:25.493 "name": "BaseBdev4", 00:29:25.493 "uuid": "069bb9e3-801f-41b8-8c73-5c2effb137f7", 00:29:25.493 "is_configured": true, 00:29:25.493 "data_offset": 0, 00:29:25.493 "data_size": 65536 00:29:25.493 } 00:29:25.493 ] 00:29:25.493 }' 00:29:25.493 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:25.494 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:25.752 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:29:25.752 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:29:25.752 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:29:25.752 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:25.752 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:25.752 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:25.752 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:25.752 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:29:25.752 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:29:25.752 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:29:25.752 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:25.752 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:25.752 [2024-11-20 13:49:32.698312] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:29:25.752 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:25.752 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:29:25.752 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:29:25.752 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:25.752 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:25.752 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:25.752 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:29:25.752 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:25.752 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:29:25.752 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:29:25.752 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:29:25.752 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:25.752 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:25.752 [2024-11-20 13:49:32.791295] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:29:26.011 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:26.011 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:29:26.011 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.012 [2024-11-20 13:49:32.876495] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:29:26.012 [2024-11-20 13:49:32.876555] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.012 BaseBdev2 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:29:26.012 13:49:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:26.012 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.012 [ 00:29:26.012 { 00:29:26.012 "name": "BaseBdev2", 00:29:26.012 "aliases": [ 00:29:26.012 "3cc65e68-4bf2-4196-bb86-40f8a168900a" 00:29:26.012 ], 00:29:26.012 "product_name": "Malloc disk", 00:29:26.012 "block_size": 512, 00:29:26.012 "num_blocks": 65536, 00:29:26.012 "uuid": "3cc65e68-4bf2-4196-bb86-40f8a168900a", 00:29:26.012 "assigned_rate_limits": { 00:29:26.012 "rw_ios_per_sec": 0, 00:29:26.012 "rw_mbytes_per_sec": 0, 00:29:26.012 "r_mbytes_per_sec": 0, 00:29:26.012 "w_mbytes_per_sec": 0 00:29:26.012 }, 00:29:26.012 "claimed": false, 00:29:26.012 "zoned": false, 00:29:26.012 "supported_io_types": { 00:29:26.012 "read": true, 00:29:26.012 "write": true, 00:29:26.012 "unmap": true, 00:29:26.012 "flush": true, 00:29:26.012 "reset": true, 00:29:26.012 "nvme_admin": false, 00:29:26.012 "nvme_io": false, 00:29:26.012 "nvme_io_md": false, 00:29:26.012 "write_zeroes": true, 00:29:26.012 "zcopy": true, 00:29:26.012 "get_zone_info": false, 00:29:26.012 "zone_management": false, 00:29:26.012 "zone_append": false, 00:29:26.012 "compare": false, 00:29:26.012 "compare_and_write": false, 00:29:26.012 "abort": true, 00:29:26.012 "seek_hole": false, 00:29:26.012 "seek_data": false, 00:29:26.012 "copy": true, 00:29:26.012 "nvme_iov_md": false 00:29:26.012 }, 00:29:26.012 "memory_domains": [ 00:29:26.012 { 00:29:26.012 "dma_device_id": "system", 00:29:26.012 "dma_device_type": 1 00:29:26.012 }, 00:29:26.012 { 00:29:26.012 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:26.012 "dma_device_type": 2 00:29:26.012 } 00:29:26.012 ], 00:29:26.012 "driver_specific": {} 00:29:26.012 } 00:29:26.012 ] 00:29:26.012 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:26.012 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:29:26.012 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:29:26.012 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:29:26.012 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:29:26.012 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:26.012 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.012 BaseBdev3 00:29:26.012 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:26.012 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:29:26.012 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:29:26.012 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:26.012 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:29:26.012 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:26.012 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:26.012 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:26.012 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:26.012 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.012 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:26.012 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:29:26.012 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:26.013 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.013 [ 00:29:26.013 { 00:29:26.013 "name": "BaseBdev3", 00:29:26.013 "aliases": [ 00:29:26.013 "4d7f1491-29e3-4fbb-bbd6-42ea40e4816c" 00:29:26.013 ], 00:29:26.013 "product_name": "Malloc disk", 00:29:26.013 "block_size": 512, 00:29:26.013 "num_blocks": 65536, 00:29:26.013 "uuid": "4d7f1491-29e3-4fbb-bbd6-42ea40e4816c", 00:29:26.013 "assigned_rate_limits": { 00:29:26.013 "rw_ios_per_sec": 0, 00:29:26.013 "rw_mbytes_per_sec": 0, 00:29:26.013 "r_mbytes_per_sec": 0, 00:29:26.013 "w_mbytes_per_sec": 0 00:29:26.013 }, 00:29:26.013 "claimed": false, 00:29:26.013 "zoned": false, 00:29:26.013 "supported_io_types": { 00:29:26.013 "read": true, 00:29:26.013 "write": true, 00:29:26.013 "unmap": true, 00:29:26.013 "flush": true, 00:29:26.013 "reset": true, 00:29:26.013 "nvme_admin": false, 00:29:26.013 "nvme_io": false, 00:29:26.013 "nvme_io_md": false, 00:29:26.013 "write_zeroes": true, 00:29:26.013 "zcopy": true, 00:29:26.013 "get_zone_info": false, 00:29:26.013 "zone_management": false, 00:29:26.013 "zone_append": false, 00:29:26.013 "compare": false, 00:29:26.013 "compare_and_write": false, 00:29:26.013 "abort": true, 00:29:26.013 "seek_hole": false, 00:29:26.013 "seek_data": false, 00:29:26.013 "copy": true, 00:29:26.013 "nvme_iov_md": false 00:29:26.013 }, 00:29:26.013 "memory_domains": [ 00:29:26.013 { 00:29:26.013 "dma_device_id": "system", 00:29:26.013 "dma_device_type": 1 00:29:26.013 }, 00:29:26.013 { 00:29:26.013 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:26.013 "dma_device_type": 2 00:29:26.013 } 00:29:26.013 ], 00:29:26.013 "driver_specific": {} 00:29:26.013 } 00:29:26.013 ] 00:29:26.013 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:26.013 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:29:26.013 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.272 BaseBdev4 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.272 [ 00:29:26.272 { 00:29:26.272 "name": "BaseBdev4", 00:29:26.272 "aliases": [ 00:29:26.272 "93dcc887-9a49-487c-82fa-52d1ef8976dd" 00:29:26.272 ], 00:29:26.272 "product_name": "Malloc disk", 00:29:26.272 "block_size": 512, 00:29:26.272 "num_blocks": 65536, 00:29:26.272 "uuid": "93dcc887-9a49-487c-82fa-52d1ef8976dd", 00:29:26.272 "assigned_rate_limits": { 00:29:26.272 "rw_ios_per_sec": 0, 00:29:26.272 "rw_mbytes_per_sec": 0, 00:29:26.272 "r_mbytes_per_sec": 0, 00:29:26.272 "w_mbytes_per_sec": 0 00:29:26.272 }, 00:29:26.272 "claimed": false, 00:29:26.272 "zoned": false, 00:29:26.272 "supported_io_types": { 00:29:26.272 "read": true, 00:29:26.272 "write": true, 00:29:26.272 "unmap": true, 00:29:26.272 "flush": true, 00:29:26.272 "reset": true, 00:29:26.272 "nvme_admin": false, 00:29:26.272 "nvme_io": false, 00:29:26.272 "nvme_io_md": false, 00:29:26.272 "write_zeroes": true, 00:29:26.272 "zcopy": true, 00:29:26.272 "get_zone_info": false, 00:29:26.272 "zone_management": false, 00:29:26.272 "zone_append": false, 00:29:26.272 "compare": false, 00:29:26.272 "compare_and_write": false, 00:29:26.272 "abort": true, 00:29:26.272 "seek_hole": false, 00:29:26.272 "seek_data": false, 00:29:26.272 "copy": true, 00:29:26.272 "nvme_iov_md": false 00:29:26.272 }, 00:29:26.272 "memory_domains": [ 00:29:26.272 { 00:29:26.272 "dma_device_id": "system", 00:29:26.272 "dma_device_type": 1 00:29:26.272 }, 00:29:26.272 { 00:29:26.272 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:26.272 "dma_device_type": 2 00:29:26.272 } 00:29:26.272 ], 00:29:26.272 "driver_specific": {} 00:29:26.272 } 00:29:26.272 ] 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.272 [2024-11-20 13:49:33.128557] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:29:26.272 [2024-11-20 13:49:33.128749] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:29:26.272 [2024-11-20 13:49:33.129163] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:29:26.272 [2024-11-20 13:49:33.130999] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:29:26.272 [2024-11-20 13:49:33.131119] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:26.272 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:26.273 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:26.273 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:26.273 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:26.273 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:26.273 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:26.273 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:26.273 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.273 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:26.273 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:26.273 "name": "Existed_Raid", 00:29:26.273 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:26.273 "strip_size_kb": 64, 00:29:26.273 "state": "configuring", 00:29:26.273 "raid_level": "concat", 00:29:26.273 "superblock": false, 00:29:26.273 "num_base_bdevs": 4, 00:29:26.273 "num_base_bdevs_discovered": 3, 00:29:26.273 "num_base_bdevs_operational": 4, 00:29:26.273 "base_bdevs_list": [ 00:29:26.273 { 00:29:26.273 "name": "BaseBdev1", 00:29:26.273 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:26.273 "is_configured": false, 00:29:26.273 "data_offset": 0, 00:29:26.273 "data_size": 0 00:29:26.273 }, 00:29:26.273 { 00:29:26.273 "name": "BaseBdev2", 00:29:26.273 "uuid": "3cc65e68-4bf2-4196-bb86-40f8a168900a", 00:29:26.273 "is_configured": true, 00:29:26.273 "data_offset": 0, 00:29:26.273 "data_size": 65536 00:29:26.273 }, 00:29:26.273 { 00:29:26.273 "name": "BaseBdev3", 00:29:26.273 "uuid": "4d7f1491-29e3-4fbb-bbd6-42ea40e4816c", 00:29:26.273 "is_configured": true, 00:29:26.273 "data_offset": 0, 00:29:26.273 "data_size": 65536 00:29:26.273 }, 00:29:26.273 { 00:29:26.273 "name": "BaseBdev4", 00:29:26.273 "uuid": "93dcc887-9a49-487c-82fa-52d1ef8976dd", 00:29:26.273 "is_configured": true, 00:29:26.273 "data_offset": 0, 00:29:26.273 "data_size": 65536 00:29:26.273 } 00:29:26.273 ] 00:29:26.273 }' 00:29:26.273 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:26.273 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.531 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:29:26.531 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:26.531 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.531 [2024-11-20 13:49:33.444636] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:29:26.531 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:26.531 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:29:26.531 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:26.531 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:26.531 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:26.531 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:26.531 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:26.531 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:26.531 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:26.531 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:26.531 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:26.531 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:26.531 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:26.531 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.531 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:26.531 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:26.531 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:26.531 "name": "Existed_Raid", 00:29:26.531 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:26.531 "strip_size_kb": 64, 00:29:26.531 "state": "configuring", 00:29:26.531 "raid_level": "concat", 00:29:26.531 "superblock": false, 00:29:26.531 "num_base_bdevs": 4, 00:29:26.531 "num_base_bdevs_discovered": 2, 00:29:26.531 "num_base_bdevs_operational": 4, 00:29:26.531 "base_bdevs_list": [ 00:29:26.531 { 00:29:26.531 "name": "BaseBdev1", 00:29:26.531 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:26.531 "is_configured": false, 00:29:26.531 "data_offset": 0, 00:29:26.531 "data_size": 0 00:29:26.531 }, 00:29:26.531 { 00:29:26.531 "name": null, 00:29:26.531 "uuid": "3cc65e68-4bf2-4196-bb86-40f8a168900a", 00:29:26.531 "is_configured": false, 00:29:26.531 "data_offset": 0, 00:29:26.531 "data_size": 65536 00:29:26.531 }, 00:29:26.531 { 00:29:26.531 "name": "BaseBdev3", 00:29:26.531 "uuid": "4d7f1491-29e3-4fbb-bbd6-42ea40e4816c", 00:29:26.531 "is_configured": true, 00:29:26.531 "data_offset": 0, 00:29:26.531 "data_size": 65536 00:29:26.531 }, 00:29:26.531 { 00:29:26.531 "name": "BaseBdev4", 00:29:26.531 "uuid": "93dcc887-9a49-487c-82fa-52d1ef8976dd", 00:29:26.531 "is_configured": true, 00:29:26.531 "data_offset": 0, 00:29:26.531 "data_size": 65536 00:29:26.531 } 00:29:26.531 ] 00:29:26.531 }' 00:29:26.531 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:26.531 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.789 [2024-11-20 13:49:33.816869] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:29:26.789 BaseBdev1 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.789 [ 00:29:26.789 { 00:29:26.789 "name": "BaseBdev1", 00:29:26.789 "aliases": [ 00:29:26.789 "fd353713-fe09-496c-a61d-dd0511c1aeee" 00:29:26.789 ], 00:29:26.789 "product_name": "Malloc disk", 00:29:26.789 "block_size": 512, 00:29:26.789 "num_blocks": 65536, 00:29:26.789 "uuid": "fd353713-fe09-496c-a61d-dd0511c1aeee", 00:29:26.789 "assigned_rate_limits": { 00:29:26.789 "rw_ios_per_sec": 0, 00:29:26.789 "rw_mbytes_per_sec": 0, 00:29:26.789 "r_mbytes_per_sec": 0, 00:29:26.789 "w_mbytes_per_sec": 0 00:29:26.789 }, 00:29:26.789 "claimed": true, 00:29:26.789 "claim_type": "exclusive_write", 00:29:26.789 "zoned": false, 00:29:26.789 "supported_io_types": { 00:29:26.789 "read": true, 00:29:26.789 "write": true, 00:29:26.789 "unmap": true, 00:29:26.789 "flush": true, 00:29:26.789 "reset": true, 00:29:26.789 "nvme_admin": false, 00:29:26.789 "nvme_io": false, 00:29:26.789 "nvme_io_md": false, 00:29:26.789 "write_zeroes": true, 00:29:26.789 "zcopy": true, 00:29:26.789 "get_zone_info": false, 00:29:26.789 "zone_management": false, 00:29:26.789 "zone_append": false, 00:29:26.789 "compare": false, 00:29:26.789 "compare_and_write": false, 00:29:26.789 "abort": true, 00:29:26.789 "seek_hole": false, 00:29:26.789 "seek_data": false, 00:29:26.789 "copy": true, 00:29:26.789 "nvme_iov_md": false 00:29:26.789 }, 00:29:26.789 "memory_domains": [ 00:29:26.789 { 00:29:26.789 "dma_device_id": "system", 00:29:26.789 "dma_device_type": 1 00:29:26.789 }, 00:29:26.789 { 00:29:26.789 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:26.789 "dma_device_type": 2 00:29:26.789 } 00:29:26.789 ], 00:29:26.789 "driver_specific": {} 00:29:26.789 } 00:29:26.789 ] 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:26.789 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:27.047 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:27.047 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:27.047 "name": "Existed_Raid", 00:29:27.047 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:27.047 "strip_size_kb": 64, 00:29:27.047 "state": "configuring", 00:29:27.047 "raid_level": "concat", 00:29:27.047 "superblock": false, 00:29:27.047 "num_base_bdevs": 4, 00:29:27.047 "num_base_bdevs_discovered": 3, 00:29:27.047 "num_base_bdevs_operational": 4, 00:29:27.047 "base_bdevs_list": [ 00:29:27.047 { 00:29:27.047 "name": "BaseBdev1", 00:29:27.047 "uuid": "fd353713-fe09-496c-a61d-dd0511c1aeee", 00:29:27.047 "is_configured": true, 00:29:27.047 "data_offset": 0, 00:29:27.047 "data_size": 65536 00:29:27.047 }, 00:29:27.047 { 00:29:27.047 "name": null, 00:29:27.047 "uuid": "3cc65e68-4bf2-4196-bb86-40f8a168900a", 00:29:27.047 "is_configured": false, 00:29:27.047 "data_offset": 0, 00:29:27.047 "data_size": 65536 00:29:27.047 }, 00:29:27.047 { 00:29:27.047 "name": "BaseBdev3", 00:29:27.047 "uuid": "4d7f1491-29e3-4fbb-bbd6-42ea40e4816c", 00:29:27.047 "is_configured": true, 00:29:27.047 "data_offset": 0, 00:29:27.047 "data_size": 65536 00:29:27.047 }, 00:29:27.047 { 00:29:27.047 "name": "BaseBdev4", 00:29:27.047 "uuid": "93dcc887-9a49-487c-82fa-52d1ef8976dd", 00:29:27.047 "is_configured": true, 00:29:27.047 "data_offset": 0, 00:29:27.047 "data_size": 65536 00:29:27.047 } 00:29:27.047 ] 00:29:27.047 }' 00:29:27.047 13:49:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:27.047 13:49:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:27.307 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:27.307 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:27.307 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:27.307 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:29:27.307 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:27.307 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:29:27.307 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:29:27.307 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:27.307 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:27.307 [2024-11-20 13:49:34.205072] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:29:27.307 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:27.307 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:29:27.307 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:27.307 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:27.307 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:27.307 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:27.307 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:27.307 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:27.307 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:27.308 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:27.308 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:27.308 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:27.308 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:27.308 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:27.308 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:27.308 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:27.308 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:27.308 "name": "Existed_Raid", 00:29:27.308 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:27.308 "strip_size_kb": 64, 00:29:27.308 "state": "configuring", 00:29:27.308 "raid_level": "concat", 00:29:27.308 "superblock": false, 00:29:27.308 "num_base_bdevs": 4, 00:29:27.308 "num_base_bdevs_discovered": 2, 00:29:27.308 "num_base_bdevs_operational": 4, 00:29:27.308 "base_bdevs_list": [ 00:29:27.308 { 00:29:27.308 "name": "BaseBdev1", 00:29:27.308 "uuid": "fd353713-fe09-496c-a61d-dd0511c1aeee", 00:29:27.308 "is_configured": true, 00:29:27.308 "data_offset": 0, 00:29:27.308 "data_size": 65536 00:29:27.308 }, 00:29:27.308 { 00:29:27.308 "name": null, 00:29:27.308 "uuid": "3cc65e68-4bf2-4196-bb86-40f8a168900a", 00:29:27.308 "is_configured": false, 00:29:27.308 "data_offset": 0, 00:29:27.308 "data_size": 65536 00:29:27.308 }, 00:29:27.308 { 00:29:27.308 "name": null, 00:29:27.308 "uuid": "4d7f1491-29e3-4fbb-bbd6-42ea40e4816c", 00:29:27.308 "is_configured": false, 00:29:27.308 "data_offset": 0, 00:29:27.308 "data_size": 65536 00:29:27.308 }, 00:29:27.308 { 00:29:27.308 "name": "BaseBdev4", 00:29:27.308 "uuid": "93dcc887-9a49-487c-82fa-52d1ef8976dd", 00:29:27.308 "is_configured": true, 00:29:27.308 "data_offset": 0, 00:29:27.308 "data_size": 65536 00:29:27.308 } 00:29:27.308 ] 00:29:27.308 }' 00:29:27.308 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:27.308 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:27.583 [2024-11-20 13:49:34.569129] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:27.583 "name": "Existed_Raid", 00:29:27.583 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:27.583 "strip_size_kb": 64, 00:29:27.583 "state": "configuring", 00:29:27.583 "raid_level": "concat", 00:29:27.583 "superblock": false, 00:29:27.583 "num_base_bdevs": 4, 00:29:27.583 "num_base_bdevs_discovered": 3, 00:29:27.583 "num_base_bdevs_operational": 4, 00:29:27.583 "base_bdevs_list": [ 00:29:27.583 { 00:29:27.583 "name": "BaseBdev1", 00:29:27.583 "uuid": "fd353713-fe09-496c-a61d-dd0511c1aeee", 00:29:27.583 "is_configured": true, 00:29:27.583 "data_offset": 0, 00:29:27.583 "data_size": 65536 00:29:27.583 }, 00:29:27.583 { 00:29:27.583 "name": null, 00:29:27.583 "uuid": "3cc65e68-4bf2-4196-bb86-40f8a168900a", 00:29:27.583 "is_configured": false, 00:29:27.583 "data_offset": 0, 00:29:27.583 "data_size": 65536 00:29:27.583 }, 00:29:27.583 { 00:29:27.583 "name": "BaseBdev3", 00:29:27.583 "uuid": "4d7f1491-29e3-4fbb-bbd6-42ea40e4816c", 00:29:27.583 "is_configured": true, 00:29:27.583 "data_offset": 0, 00:29:27.583 "data_size": 65536 00:29:27.583 }, 00:29:27.583 { 00:29:27.583 "name": "BaseBdev4", 00:29:27.583 "uuid": "93dcc887-9a49-487c-82fa-52d1ef8976dd", 00:29:27.583 "is_configured": true, 00:29:27.583 "data_offset": 0, 00:29:27.583 "data_size": 65536 00:29:27.583 } 00:29:27.583 ] 00:29:27.583 }' 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:27.583 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:27.840 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:29:27.840 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:27.840 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:27.840 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:28.098 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:28.098 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:29:28.098 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:29:28.098 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:28.098 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:28.098 [2024-11-20 13:49:34.929248] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:29:28.098 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:28.098 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:29:28.098 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:28.098 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:28.098 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:28.098 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:28.098 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:28.098 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:28.098 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:28.098 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:28.098 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:28.098 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:28.098 13:49:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:28.098 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:28.098 13:49:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:28.098 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:28.098 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:28.098 "name": "Existed_Raid", 00:29:28.098 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:28.098 "strip_size_kb": 64, 00:29:28.098 "state": "configuring", 00:29:28.098 "raid_level": "concat", 00:29:28.098 "superblock": false, 00:29:28.098 "num_base_bdevs": 4, 00:29:28.098 "num_base_bdevs_discovered": 2, 00:29:28.098 "num_base_bdevs_operational": 4, 00:29:28.098 "base_bdevs_list": [ 00:29:28.098 { 00:29:28.098 "name": null, 00:29:28.098 "uuid": "fd353713-fe09-496c-a61d-dd0511c1aeee", 00:29:28.098 "is_configured": false, 00:29:28.098 "data_offset": 0, 00:29:28.098 "data_size": 65536 00:29:28.098 }, 00:29:28.098 { 00:29:28.098 "name": null, 00:29:28.098 "uuid": "3cc65e68-4bf2-4196-bb86-40f8a168900a", 00:29:28.098 "is_configured": false, 00:29:28.098 "data_offset": 0, 00:29:28.098 "data_size": 65536 00:29:28.098 }, 00:29:28.098 { 00:29:28.098 "name": "BaseBdev3", 00:29:28.098 "uuid": "4d7f1491-29e3-4fbb-bbd6-42ea40e4816c", 00:29:28.098 "is_configured": true, 00:29:28.098 "data_offset": 0, 00:29:28.098 "data_size": 65536 00:29:28.098 }, 00:29:28.098 { 00:29:28.098 "name": "BaseBdev4", 00:29:28.098 "uuid": "93dcc887-9a49-487c-82fa-52d1ef8976dd", 00:29:28.098 "is_configured": true, 00:29:28.098 "data_offset": 0, 00:29:28.098 "data_size": 65536 00:29:28.098 } 00:29:28.098 ] 00:29:28.098 }' 00:29:28.098 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:28.098 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:28.356 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:28.356 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:28.356 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:28.356 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:29:28.356 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:28.356 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:29:28.356 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:29:28.356 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:28.356 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:28.356 [2024-11-20 13:49:35.343354] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:29:28.356 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:28.356 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:29:28.356 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:28.356 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:28.356 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:28.356 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:28.356 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:28.356 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:28.356 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:28.356 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:28.356 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:28.356 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:28.356 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:28.356 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:28.357 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:28.357 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:28.357 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:28.357 "name": "Existed_Raid", 00:29:28.357 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:28.357 "strip_size_kb": 64, 00:29:28.357 "state": "configuring", 00:29:28.357 "raid_level": "concat", 00:29:28.357 "superblock": false, 00:29:28.357 "num_base_bdevs": 4, 00:29:28.357 "num_base_bdevs_discovered": 3, 00:29:28.357 "num_base_bdevs_operational": 4, 00:29:28.357 "base_bdevs_list": [ 00:29:28.357 { 00:29:28.357 "name": null, 00:29:28.357 "uuid": "fd353713-fe09-496c-a61d-dd0511c1aeee", 00:29:28.357 "is_configured": false, 00:29:28.357 "data_offset": 0, 00:29:28.357 "data_size": 65536 00:29:28.357 }, 00:29:28.357 { 00:29:28.357 "name": "BaseBdev2", 00:29:28.357 "uuid": "3cc65e68-4bf2-4196-bb86-40f8a168900a", 00:29:28.357 "is_configured": true, 00:29:28.357 "data_offset": 0, 00:29:28.357 "data_size": 65536 00:29:28.357 }, 00:29:28.357 { 00:29:28.357 "name": "BaseBdev3", 00:29:28.357 "uuid": "4d7f1491-29e3-4fbb-bbd6-42ea40e4816c", 00:29:28.357 "is_configured": true, 00:29:28.357 "data_offset": 0, 00:29:28.357 "data_size": 65536 00:29:28.357 }, 00:29:28.357 { 00:29:28.357 "name": "BaseBdev4", 00:29:28.357 "uuid": "93dcc887-9a49-487c-82fa-52d1ef8976dd", 00:29:28.357 "is_configured": true, 00:29:28.357 "data_offset": 0, 00:29:28.357 "data_size": 65536 00:29:28.357 } 00:29:28.357 ] 00:29:28.357 }' 00:29:28.357 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:28.357 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:28.614 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:28.614 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:28.614 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:28.614 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:29:28.873 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:28.873 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:29:28.873 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:29:28.873 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:28.873 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:28.873 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:28.873 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:28.873 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u fd353713-fe09-496c-a61d-dd0511c1aeee 00:29:28.873 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:28.873 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:28.873 [2024-11-20 13:49:35.748261] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:29:28.873 [2024-11-20 13:49:35.748329] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:29:28.873 [2024-11-20 13:49:35.748335] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:29:28.873 [2024-11-20 13:49:35.748584] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:29:28.873 [2024-11-20 13:49:35.748706] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:29:28.873 [2024-11-20 13:49:35.748715] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:29:28.873 [2024-11-20 13:49:35.748926] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:28.873 NewBaseBdev 00:29:28.873 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:28.873 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:29:28.873 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:29:28.873 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:28.873 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:29:28.873 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:28.873 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:28.873 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:28.873 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:28.873 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:28.873 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:28.873 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:29:28.873 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:28.873 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:28.873 [ 00:29:28.873 { 00:29:28.873 "name": "NewBaseBdev", 00:29:28.873 "aliases": [ 00:29:28.873 "fd353713-fe09-496c-a61d-dd0511c1aeee" 00:29:28.873 ], 00:29:28.873 "product_name": "Malloc disk", 00:29:28.873 "block_size": 512, 00:29:28.873 "num_blocks": 65536, 00:29:28.873 "uuid": "fd353713-fe09-496c-a61d-dd0511c1aeee", 00:29:28.873 "assigned_rate_limits": { 00:29:28.873 "rw_ios_per_sec": 0, 00:29:28.873 "rw_mbytes_per_sec": 0, 00:29:28.873 "r_mbytes_per_sec": 0, 00:29:28.873 "w_mbytes_per_sec": 0 00:29:28.873 }, 00:29:28.873 "claimed": true, 00:29:28.873 "claim_type": "exclusive_write", 00:29:28.873 "zoned": false, 00:29:28.873 "supported_io_types": { 00:29:28.873 "read": true, 00:29:28.873 "write": true, 00:29:28.873 "unmap": true, 00:29:28.873 "flush": true, 00:29:28.873 "reset": true, 00:29:28.873 "nvme_admin": false, 00:29:28.873 "nvme_io": false, 00:29:28.873 "nvme_io_md": false, 00:29:28.873 "write_zeroes": true, 00:29:28.873 "zcopy": true, 00:29:28.873 "get_zone_info": false, 00:29:28.873 "zone_management": false, 00:29:28.873 "zone_append": false, 00:29:28.873 "compare": false, 00:29:28.873 "compare_and_write": false, 00:29:28.873 "abort": true, 00:29:28.873 "seek_hole": false, 00:29:28.873 "seek_data": false, 00:29:28.873 "copy": true, 00:29:28.873 "nvme_iov_md": false 00:29:28.873 }, 00:29:28.873 "memory_domains": [ 00:29:28.873 { 00:29:28.873 "dma_device_id": "system", 00:29:28.873 "dma_device_type": 1 00:29:28.873 }, 00:29:28.873 { 00:29:28.874 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:28.874 "dma_device_type": 2 00:29:28.874 } 00:29:28.874 ], 00:29:28.874 "driver_specific": {} 00:29:28.874 } 00:29:28.874 ] 00:29:28.874 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:28.874 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:29:28.874 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:29:28.874 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:28.874 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:28.874 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:28.874 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:28.874 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:28.874 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:28.874 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:28.874 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:28.874 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:28.874 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:28.874 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:28.874 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:28.874 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:28.874 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:28.874 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:28.874 "name": "Existed_Raid", 00:29:28.874 "uuid": "29b58d5d-fcd4-4611-98ca-4c89e8ba5033", 00:29:28.874 "strip_size_kb": 64, 00:29:28.874 "state": "online", 00:29:28.874 "raid_level": "concat", 00:29:28.874 "superblock": false, 00:29:28.874 "num_base_bdevs": 4, 00:29:28.874 "num_base_bdevs_discovered": 4, 00:29:28.874 "num_base_bdevs_operational": 4, 00:29:28.874 "base_bdevs_list": [ 00:29:28.874 { 00:29:28.874 "name": "NewBaseBdev", 00:29:28.874 "uuid": "fd353713-fe09-496c-a61d-dd0511c1aeee", 00:29:28.874 "is_configured": true, 00:29:28.874 "data_offset": 0, 00:29:28.874 "data_size": 65536 00:29:28.874 }, 00:29:28.874 { 00:29:28.874 "name": "BaseBdev2", 00:29:28.874 "uuid": "3cc65e68-4bf2-4196-bb86-40f8a168900a", 00:29:28.874 "is_configured": true, 00:29:28.874 "data_offset": 0, 00:29:28.874 "data_size": 65536 00:29:28.874 }, 00:29:28.874 { 00:29:28.874 "name": "BaseBdev3", 00:29:28.874 "uuid": "4d7f1491-29e3-4fbb-bbd6-42ea40e4816c", 00:29:28.874 "is_configured": true, 00:29:28.874 "data_offset": 0, 00:29:28.874 "data_size": 65536 00:29:28.874 }, 00:29:28.874 { 00:29:28.874 "name": "BaseBdev4", 00:29:28.874 "uuid": "93dcc887-9a49-487c-82fa-52d1ef8976dd", 00:29:28.874 "is_configured": true, 00:29:28.874 "data_offset": 0, 00:29:28.874 "data_size": 65536 00:29:28.874 } 00:29:28.874 ] 00:29:28.874 }' 00:29:28.874 13:49:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:28.874 13:49:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:29.133 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:29:29.133 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:29:29.133 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:29:29.133 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:29:29.133 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:29:29.133 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:29:29.133 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:29:29.133 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:29:29.133 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:29.133 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:29.133 [2024-11-20 13:49:36.144784] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:29.133 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:29.133 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:29.133 "name": "Existed_Raid", 00:29:29.133 "aliases": [ 00:29:29.133 "29b58d5d-fcd4-4611-98ca-4c89e8ba5033" 00:29:29.133 ], 00:29:29.133 "product_name": "Raid Volume", 00:29:29.133 "block_size": 512, 00:29:29.133 "num_blocks": 262144, 00:29:29.133 "uuid": "29b58d5d-fcd4-4611-98ca-4c89e8ba5033", 00:29:29.133 "assigned_rate_limits": { 00:29:29.133 "rw_ios_per_sec": 0, 00:29:29.133 "rw_mbytes_per_sec": 0, 00:29:29.133 "r_mbytes_per_sec": 0, 00:29:29.133 "w_mbytes_per_sec": 0 00:29:29.133 }, 00:29:29.133 "claimed": false, 00:29:29.133 "zoned": false, 00:29:29.133 "supported_io_types": { 00:29:29.133 "read": true, 00:29:29.133 "write": true, 00:29:29.133 "unmap": true, 00:29:29.133 "flush": true, 00:29:29.133 "reset": true, 00:29:29.133 "nvme_admin": false, 00:29:29.133 "nvme_io": false, 00:29:29.133 "nvme_io_md": false, 00:29:29.133 "write_zeroes": true, 00:29:29.133 "zcopy": false, 00:29:29.133 "get_zone_info": false, 00:29:29.133 "zone_management": false, 00:29:29.133 "zone_append": false, 00:29:29.133 "compare": false, 00:29:29.133 "compare_and_write": false, 00:29:29.133 "abort": false, 00:29:29.133 "seek_hole": false, 00:29:29.133 "seek_data": false, 00:29:29.133 "copy": false, 00:29:29.133 "nvme_iov_md": false 00:29:29.133 }, 00:29:29.133 "memory_domains": [ 00:29:29.133 { 00:29:29.133 "dma_device_id": "system", 00:29:29.133 "dma_device_type": 1 00:29:29.133 }, 00:29:29.133 { 00:29:29.133 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:29.133 "dma_device_type": 2 00:29:29.133 }, 00:29:29.133 { 00:29:29.133 "dma_device_id": "system", 00:29:29.133 "dma_device_type": 1 00:29:29.133 }, 00:29:29.133 { 00:29:29.133 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:29.133 "dma_device_type": 2 00:29:29.133 }, 00:29:29.133 { 00:29:29.133 "dma_device_id": "system", 00:29:29.133 "dma_device_type": 1 00:29:29.133 }, 00:29:29.133 { 00:29:29.133 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:29.133 "dma_device_type": 2 00:29:29.133 }, 00:29:29.133 { 00:29:29.133 "dma_device_id": "system", 00:29:29.133 "dma_device_type": 1 00:29:29.133 }, 00:29:29.133 { 00:29:29.133 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:29.133 "dma_device_type": 2 00:29:29.133 } 00:29:29.133 ], 00:29:29.133 "driver_specific": { 00:29:29.133 "raid": { 00:29:29.133 "uuid": "29b58d5d-fcd4-4611-98ca-4c89e8ba5033", 00:29:29.133 "strip_size_kb": 64, 00:29:29.133 "state": "online", 00:29:29.133 "raid_level": "concat", 00:29:29.133 "superblock": false, 00:29:29.133 "num_base_bdevs": 4, 00:29:29.133 "num_base_bdevs_discovered": 4, 00:29:29.133 "num_base_bdevs_operational": 4, 00:29:29.133 "base_bdevs_list": [ 00:29:29.133 { 00:29:29.133 "name": "NewBaseBdev", 00:29:29.133 "uuid": "fd353713-fe09-496c-a61d-dd0511c1aeee", 00:29:29.133 "is_configured": true, 00:29:29.133 "data_offset": 0, 00:29:29.133 "data_size": 65536 00:29:29.133 }, 00:29:29.133 { 00:29:29.133 "name": "BaseBdev2", 00:29:29.133 "uuid": "3cc65e68-4bf2-4196-bb86-40f8a168900a", 00:29:29.133 "is_configured": true, 00:29:29.133 "data_offset": 0, 00:29:29.133 "data_size": 65536 00:29:29.133 }, 00:29:29.133 { 00:29:29.133 "name": "BaseBdev3", 00:29:29.133 "uuid": "4d7f1491-29e3-4fbb-bbd6-42ea40e4816c", 00:29:29.133 "is_configured": true, 00:29:29.133 "data_offset": 0, 00:29:29.133 "data_size": 65536 00:29:29.133 }, 00:29:29.133 { 00:29:29.133 "name": "BaseBdev4", 00:29:29.133 "uuid": "93dcc887-9a49-487c-82fa-52d1ef8976dd", 00:29:29.133 "is_configured": true, 00:29:29.133 "data_offset": 0, 00:29:29.133 "data_size": 65536 00:29:29.133 } 00:29:29.133 ] 00:29:29.133 } 00:29:29.133 } 00:29:29.133 }' 00:29:29.133 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:29:29.392 BaseBdev2 00:29:29.392 BaseBdev3 00:29:29.392 BaseBdev4' 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:29.392 [2024-11-20 13:49:36.384454] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:29:29.392 [2024-11-20 13:49:36.384586] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:29:29.392 [2024-11-20 13:49:36.384711] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:29.392 [2024-11-20 13:49:36.384823] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:29:29.392 [2024-11-20 13:49:36.384880] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 69411 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 69411 ']' 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 69411 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69411 00:29:29.392 killing process with pid 69411 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69411' 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 69411 00:29:29.392 [2024-11-20 13:49:36.414553] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:29:29.392 13:49:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 69411 00:29:29.650 [2024-11-20 13:49:36.626880] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:29:30.214 ************************************ 00:29:30.214 END TEST raid_state_function_test 00:29:30.214 ************************************ 00:29:30.214 13:49:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:29:30.214 00:29:30.214 real 0m8.259s 00:29:30.214 user 0m13.207s 00:29:30.214 sys 0m1.366s 00:29:30.214 13:49:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:29:30.214 13:49:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:30.473 13:49:37 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 4 true 00:29:30.473 13:49:37 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:29:30.473 13:49:37 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:29:30.473 13:49:37 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:29:30.473 ************************************ 00:29:30.473 START TEST raid_state_function_test_sb 00:29:30.473 ************************************ 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 4 true 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:29:30.473 Process raid pid: 70049 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=70049 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 70049' 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 70049 00:29:30.473 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 70049 ']' 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:29:30.473 13:49:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:30.473 [2024-11-20 13:49:37.394152] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:29:30.473 [2024-11-20 13:49:37.395039] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:29:30.731 [2024-11-20 13:49:37.558068] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:30.731 [2024-11-20 13:49:37.676788] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:29:30.989 [2024-11-20 13:49:37.827023] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:30.989 [2024-11-20 13:49:37.827070] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:31.257 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:29:31.257 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:29:31.257 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:29:31.257 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:31.257 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:31.257 [2024-11-20 13:49:38.270183] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:29:31.257 [2024-11-20 13:49:38.270406] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:29:31.257 [2024-11-20 13:49:38.270428] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:29:31.257 [2024-11-20 13:49:38.270438] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:29:31.257 [2024-11-20 13:49:38.270446] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:29:31.257 [2024-11-20 13:49:38.270456] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:29:31.257 [2024-11-20 13:49:38.270462] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:29:31.257 [2024-11-20 13:49:38.270471] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:29:31.257 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:31.257 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:29:31.257 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:31.257 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:31.257 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:31.257 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:31.257 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:31.257 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:31.257 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:31.257 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:31.257 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:31.257 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:31.257 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:31.257 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:31.257 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:31.257 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:31.257 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:31.257 "name": "Existed_Raid", 00:29:31.257 "uuid": "0e4b260e-08fd-4224-b468-1246e39f7c01", 00:29:31.257 "strip_size_kb": 64, 00:29:31.257 "state": "configuring", 00:29:31.257 "raid_level": "concat", 00:29:31.257 "superblock": true, 00:29:31.257 "num_base_bdevs": 4, 00:29:31.257 "num_base_bdevs_discovered": 0, 00:29:31.257 "num_base_bdevs_operational": 4, 00:29:31.257 "base_bdevs_list": [ 00:29:31.257 { 00:29:31.257 "name": "BaseBdev1", 00:29:31.257 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:31.257 "is_configured": false, 00:29:31.257 "data_offset": 0, 00:29:31.257 "data_size": 0 00:29:31.257 }, 00:29:31.257 { 00:29:31.257 "name": "BaseBdev2", 00:29:31.257 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:31.257 "is_configured": false, 00:29:31.257 "data_offset": 0, 00:29:31.257 "data_size": 0 00:29:31.257 }, 00:29:31.257 { 00:29:31.257 "name": "BaseBdev3", 00:29:31.257 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:31.257 "is_configured": false, 00:29:31.257 "data_offset": 0, 00:29:31.257 "data_size": 0 00:29:31.257 }, 00:29:31.257 { 00:29:31.257 "name": "BaseBdev4", 00:29:31.257 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:31.257 "is_configured": false, 00:29:31.257 "data_offset": 0, 00:29:31.257 "data_size": 0 00:29:31.257 } 00:29:31.257 ] 00:29:31.257 }' 00:29:31.257 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:31.258 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:31.824 [2024-11-20 13:49:38.586193] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:29:31.824 [2024-11-20 13:49:38.586238] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:31.824 [2024-11-20 13:49:38.594187] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:29:31.824 [2024-11-20 13:49:38.594230] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:29:31.824 [2024-11-20 13:49:38.594239] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:29:31.824 [2024-11-20 13:49:38.594248] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:29:31.824 [2024-11-20 13:49:38.594254] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:29:31.824 [2024-11-20 13:49:38.594263] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:29:31.824 [2024-11-20 13:49:38.594269] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:29:31.824 [2024-11-20 13:49:38.594278] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:31.824 BaseBdev1 00:29:31.824 [2024-11-20 13:49:38.628751] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:31.824 [ 00:29:31.824 { 00:29:31.824 "name": "BaseBdev1", 00:29:31.824 "aliases": [ 00:29:31.824 "7783341b-483f-42da-a0ff-6443837b16bc" 00:29:31.824 ], 00:29:31.824 "product_name": "Malloc disk", 00:29:31.824 "block_size": 512, 00:29:31.824 "num_blocks": 65536, 00:29:31.824 "uuid": "7783341b-483f-42da-a0ff-6443837b16bc", 00:29:31.824 "assigned_rate_limits": { 00:29:31.824 "rw_ios_per_sec": 0, 00:29:31.824 "rw_mbytes_per_sec": 0, 00:29:31.824 "r_mbytes_per_sec": 0, 00:29:31.824 "w_mbytes_per_sec": 0 00:29:31.824 }, 00:29:31.824 "claimed": true, 00:29:31.824 "claim_type": "exclusive_write", 00:29:31.824 "zoned": false, 00:29:31.824 "supported_io_types": { 00:29:31.824 "read": true, 00:29:31.824 "write": true, 00:29:31.824 "unmap": true, 00:29:31.824 "flush": true, 00:29:31.824 "reset": true, 00:29:31.824 "nvme_admin": false, 00:29:31.824 "nvme_io": false, 00:29:31.824 "nvme_io_md": false, 00:29:31.824 "write_zeroes": true, 00:29:31.824 "zcopy": true, 00:29:31.824 "get_zone_info": false, 00:29:31.824 "zone_management": false, 00:29:31.824 "zone_append": false, 00:29:31.824 "compare": false, 00:29:31.824 "compare_and_write": false, 00:29:31.824 "abort": true, 00:29:31.824 "seek_hole": false, 00:29:31.824 "seek_data": false, 00:29:31.824 "copy": true, 00:29:31.824 "nvme_iov_md": false 00:29:31.824 }, 00:29:31.824 "memory_domains": [ 00:29:31.824 { 00:29:31.824 "dma_device_id": "system", 00:29:31.824 "dma_device_type": 1 00:29:31.824 }, 00:29:31.824 { 00:29:31.824 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:31.824 "dma_device_type": 2 00:29:31.824 } 00:29:31.824 ], 00:29:31.824 "driver_specific": {} 00:29:31.824 } 00:29:31.824 ] 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:31.824 "name": "Existed_Raid", 00:29:31.824 "uuid": "f0667b41-2175-44e8-97c8-1ddca026e0f3", 00:29:31.824 "strip_size_kb": 64, 00:29:31.824 "state": "configuring", 00:29:31.824 "raid_level": "concat", 00:29:31.824 "superblock": true, 00:29:31.824 "num_base_bdevs": 4, 00:29:31.824 "num_base_bdevs_discovered": 1, 00:29:31.824 "num_base_bdevs_operational": 4, 00:29:31.824 "base_bdevs_list": [ 00:29:31.824 { 00:29:31.824 "name": "BaseBdev1", 00:29:31.824 "uuid": "7783341b-483f-42da-a0ff-6443837b16bc", 00:29:31.824 "is_configured": true, 00:29:31.824 "data_offset": 2048, 00:29:31.824 "data_size": 63488 00:29:31.824 }, 00:29:31.824 { 00:29:31.824 "name": "BaseBdev2", 00:29:31.824 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:31.824 "is_configured": false, 00:29:31.824 "data_offset": 0, 00:29:31.824 "data_size": 0 00:29:31.824 }, 00:29:31.824 { 00:29:31.824 "name": "BaseBdev3", 00:29:31.824 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:31.824 "is_configured": false, 00:29:31.824 "data_offset": 0, 00:29:31.824 "data_size": 0 00:29:31.824 }, 00:29:31.824 { 00:29:31.824 "name": "BaseBdev4", 00:29:31.824 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:31.824 "is_configured": false, 00:29:31.824 "data_offset": 0, 00:29:31.824 "data_size": 0 00:29:31.824 } 00:29:31.824 ] 00:29:31.824 }' 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:31.824 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:32.083 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:29:32.083 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:32.083 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:32.083 [2024-11-20 13:49:38.968920] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:29:32.083 [2024-11-20 13:49:38.969000] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:29:32.083 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:32.083 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:29:32.083 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:32.083 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:32.083 [2024-11-20 13:49:38.976958] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:29:32.083 [2024-11-20 13:49:38.978968] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:29:32.083 [2024-11-20 13:49:38.979029] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:29:32.083 [2024-11-20 13:49:38.979038] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:29:32.083 [2024-11-20 13:49:38.979050] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:29:32.083 [2024-11-20 13:49:38.979057] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:29:32.083 [2024-11-20 13:49:38.979066] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:29:32.083 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:32.083 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:29:32.083 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:29:32.083 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:29:32.083 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:32.083 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:32.083 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:32.083 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:32.083 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:32.083 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:32.083 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:32.083 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:32.083 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:32.083 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:32.083 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:32.083 13:49:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:32.083 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:32.083 13:49:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:32.083 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:32.083 "name": "Existed_Raid", 00:29:32.083 "uuid": "43b47966-e78f-49ee-980d-85c2383911d1", 00:29:32.083 "strip_size_kb": 64, 00:29:32.083 "state": "configuring", 00:29:32.083 "raid_level": "concat", 00:29:32.083 "superblock": true, 00:29:32.083 "num_base_bdevs": 4, 00:29:32.083 "num_base_bdevs_discovered": 1, 00:29:32.083 "num_base_bdevs_operational": 4, 00:29:32.083 "base_bdevs_list": [ 00:29:32.083 { 00:29:32.083 "name": "BaseBdev1", 00:29:32.083 "uuid": "7783341b-483f-42da-a0ff-6443837b16bc", 00:29:32.083 "is_configured": true, 00:29:32.083 "data_offset": 2048, 00:29:32.083 "data_size": 63488 00:29:32.083 }, 00:29:32.083 { 00:29:32.083 "name": "BaseBdev2", 00:29:32.083 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:32.083 "is_configured": false, 00:29:32.083 "data_offset": 0, 00:29:32.083 "data_size": 0 00:29:32.083 }, 00:29:32.083 { 00:29:32.083 "name": "BaseBdev3", 00:29:32.083 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:32.083 "is_configured": false, 00:29:32.083 "data_offset": 0, 00:29:32.083 "data_size": 0 00:29:32.083 }, 00:29:32.083 { 00:29:32.084 "name": "BaseBdev4", 00:29:32.084 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:32.084 "is_configured": false, 00:29:32.084 "data_offset": 0, 00:29:32.084 "data_size": 0 00:29:32.084 } 00:29:32.084 ] 00:29:32.084 }' 00:29:32.084 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:32.084 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:32.342 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:29:32.342 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:32.342 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:32.342 [2024-11-20 13:49:39.298086] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:29:32.342 BaseBdev2 00:29:32.342 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:32.342 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:29:32.342 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:29:32.342 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:32.342 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:29:32.342 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:32.342 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:32.342 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:32.342 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:32.342 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:32.342 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:32.342 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:29:32.342 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:32.342 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:32.342 [ 00:29:32.342 { 00:29:32.342 "name": "BaseBdev2", 00:29:32.342 "aliases": [ 00:29:32.342 "a6b73762-4a53-4bac-8d7f-66c0efc1e86f" 00:29:32.342 ], 00:29:32.342 "product_name": "Malloc disk", 00:29:32.342 "block_size": 512, 00:29:32.342 "num_blocks": 65536, 00:29:32.342 "uuid": "a6b73762-4a53-4bac-8d7f-66c0efc1e86f", 00:29:32.342 "assigned_rate_limits": { 00:29:32.342 "rw_ios_per_sec": 0, 00:29:32.342 "rw_mbytes_per_sec": 0, 00:29:32.342 "r_mbytes_per_sec": 0, 00:29:32.342 "w_mbytes_per_sec": 0 00:29:32.342 }, 00:29:32.342 "claimed": true, 00:29:32.342 "claim_type": "exclusive_write", 00:29:32.342 "zoned": false, 00:29:32.342 "supported_io_types": { 00:29:32.342 "read": true, 00:29:32.342 "write": true, 00:29:32.342 "unmap": true, 00:29:32.342 "flush": true, 00:29:32.342 "reset": true, 00:29:32.342 "nvme_admin": false, 00:29:32.342 "nvme_io": false, 00:29:32.342 "nvme_io_md": false, 00:29:32.342 "write_zeroes": true, 00:29:32.342 "zcopy": true, 00:29:32.342 "get_zone_info": false, 00:29:32.342 "zone_management": false, 00:29:32.342 "zone_append": false, 00:29:32.342 "compare": false, 00:29:32.342 "compare_and_write": false, 00:29:32.342 "abort": true, 00:29:32.342 "seek_hole": false, 00:29:32.342 "seek_data": false, 00:29:32.342 "copy": true, 00:29:32.342 "nvme_iov_md": false 00:29:32.342 }, 00:29:32.342 "memory_domains": [ 00:29:32.342 { 00:29:32.342 "dma_device_id": "system", 00:29:32.342 "dma_device_type": 1 00:29:32.342 }, 00:29:32.342 { 00:29:32.342 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:32.342 "dma_device_type": 2 00:29:32.342 } 00:29:32.342 ], 00:29:32.342 "driver_specific": {} 00:29:32.342 } 00:29:32.342 ] 00:29:32.342 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:32.342 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:29:32.342 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:29:32.342 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:29:32.342 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:29:32.342 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:32.342 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:32.342 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:32.342 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:32.343 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:32.343 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:32.343 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:32.343 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:32.343 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:32.343 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:32.343 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:32.343 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:32.343 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:32.343 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:32.343 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:32.343 "name": "Existed_Raid", 00:29:32.343 "uuid": "43b47966-e78f-49ee-980d-85c2383911d1", 00:29:32.343 "strip_size_kb": 64, 00:29:32.343 "state": "configuring", 00:29:32.343 "raid_level": "concat", 00:29:32.343 "superblock": true, 00:29:32.343 "num_base_bdevs": 4, 00:29:32.343 "num_base_bdevs_discovered": 2, 00:29:32.343 "num_base_bdevs_operational": 4, 00:29:32.343 "base_bdevs_list": [ 00:29:32.343 { 00:29:32.343 "name": "BaseBdev1", 00:29:32.343 "uuid": "7783341b-483f-42da-a0ff-6443837b16bc", 00:29:32.343 "is_configured": true, 00:29:32.343 "data_offset": 2048, 00:29:32.343 "data_size": 63488 00:29:32.343 }, 00:29:32.343 { 00:29:32.343 "name": "BaseBdev2", 00:29:32.343 "uuid": "a6b73762-4a53-4bac-8d7f-66c0efc1e86f", 00:29:32.343 "is_configured": true, 00:29:32.343 "data_offset": 2048, 00:29:32.343 "data_size": 63488 00:29:32.343 }, 00:29:32.343 { 00:29:32.343 "name": "BaseBdev3", 00:29:32.343 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:32.343 "is_configured": false, 00:29:32.343 "data_offset": 0, 00:29:32.343 "data_size": 0 00:29:32.343 }, 00:29:32.343 { 00:29:32.343 "name": "BaseBdev4", 00:29:32.343 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:32.343 "is_configured": false, 00:29:32.343 "data_offset": 0, 00:29:32.343 "data_size": 0 00:29:32.343 } 00:29:32.343 ] 00:29:32.343 }' 00:29:32.343 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:32.343 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:32.909 [2024-11-20 13:49:39.724269] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:29:32.909 BaseBdev3 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:32.909 [ 00:29:32.909 { 00:29:32.909 "name": "BaseBdev3", 00:29:32.909 "aliases": [ 00:29:32.909 "b5476039-38ee-4d53-a1a9-d0438b021262" 00:29:32.909 ], 00:29:32.909 "product_name": "Malloc disk", 00:29:32.909 "block_size": 512, 00:29:32.909 "num_blocks": 65536, 00:29:32.909 "uuid": "b5476039-38ee-4d53-a1a9-d0438b021262", 00:29:32.909 "assigned_rate_limits": { 00:29:32.909 "rw_ios_per_sec": 0, 00:29:32.909 "rw_mbytes_per_sec": 0, 00:29:32.909 "r_mbytes_per_sec": 0, 00:29:32.909 "w_mbytes_per_sec": 0 00:29:32.909 }, 00:29:32.909 "claimed": true, 00:29:32.909 "claim_type": "exclusive_write", 00:29:32.909 "zoned": false, 00:29:32.909 "supported_io_types": { 00:29:32.909 "read": true, 00:29:32.909 "write": true, 00:29:32.909 "unmap": true, 00:29:32.909 "flush": true, 00:29:32.909 "reset": true, 00:29:32.909 "nvme_admin": false, 00:29:32.909 "nvme_io": false, 00:29:32.909 "nvme_io_md": false, 00:29:32.909 "write_zeroes": true, 00:29:32.909 "zcopy": true, 00:29:32.909 "get_zone_info": false, 00:29:32.909 "zone_management": false, 00:29:32.909 "zone_append": false, 00:29:32.909 "compare": false, 00:29:32.909 "compare_and_write": false, 00:29:32.909 "abort": true, 00:29:32.909 "seek_hole": false, 00:29:32.909 "seek_data": false, 00:29:32.909 "copy": true, 00:29:32.909 "nvme_iov_md": false 00:29:32.909 }, 00:29:32.909 "memory_domains": [ 00:29:32.909 { 00:29:32.909 "dma_device_id": "system", 00:29:32.909 "dma_device_type": 1 00:29:32.909 }, 00:29:32.909 { 00:29:32.909 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:32.909 "dma_device_type": 2 00:29:32.909 } 00:29:32.909 ], 00:29:32.909 "driver_specific": {} 00:29:32.909 } 00:29:32.909 ] 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:32.909 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:32.909 "name": "Existed_Raid", 00:29:32.909 "uuid": "43b47966-e78f-49ee-980d-85c2383911d1", 00:29:32.909 "strip_size_kb": 64, 00:29:32.909 "state": "configuring", 00:29:32.909 "raid_level": "concat", 00:29:32.909 "superblock": true, 00:29:32.909 "num_base_bdevs": 4, 00:29:32.909 "num_base_bdevs_discovered": 3, 00:29:32.909 "num_base_bdevs_operational": 4, 00:29:32.909 "base_bdevs_list": [ 00:29:32.909 { 00:29:32.909 "name": "BaseBdev1", 00:29:32.909 "uuid": "7783341b-483f-42da-a0ff-6443837b16bc", 00:29:32.909 "is_configured": true, 00:29:32.909 "data_offset": 2048, 00:29:32.909 "data_size": 63488 00:29:32.909 }, 00:29:32.909 { 00:29:32.909 "name": "BaseBdev2", 00:29:32.909 "uuid": "a6b73762-4a53-4bac-8d7f-66c0efc1e86f", 00:29:32.909 "is_configured": true, 00:29:32.909 "data_offset": 2048, 00:29:32.909 "data_size": 63488 00:29:32.909 }, 00:29:32.909 { 00:29:32.909 "name": "BaseBdev3", 00:29:32.910 "uuid": "b5476039-38ee-4d53-a1a9-d0438b021262", 00:29:32.910 "is_configured": true, 00:29:32.910 "data_offset": 2048, 00:29:32.910 "data_size": 63488 00:29:32.910 }, 00:29:32.910 { 00:29:32.910 "name": "BaseBdev4", 00:29:32.910 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:32.910 "is_configured": false, 00:29:32.910 "data_offset": 0, 00:29:32.910 "data_size": 0 00:29:32.910 } 00:29:32.910 ] 00:29:32.910 }' 00:29:32.910 13:49:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:32.910 13:49:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:33.168 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:29:33.168 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:33.168 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:33.168 [2024-11-20 13:49:40.081259] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:29:33.168 BaseBdev4 00:29:33.168 [2024-11-20 13:49:40.081683] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:29:33.168 [2024-11-20 13:49:40.081703] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:29:33.168 [2024-11-20 13:49:40.082017] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:29:33.168 [2024-11-20 13:49:40.082158] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:29:33.168 [2024-11-20 13:49:40.082170] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:29:33.168 [2024-11-20 13:49:40.082312] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:33.168 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:33.168 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:29:33.168 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:29:33.168 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:33.168 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:29:33.168 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:33.168 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:33.168 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:33.168 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:33.168 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:33.168 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:33.168 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:29:33.168 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:33.168 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:33.168 [ 00:29:33.168 { 00:29:33.168 "name": "BaseBdev4", 00:29:33.168 "aliases": [ 00:29:33.168 "b79b52eb-a370-4673-b982-de7baf01575b" 00:29:33.168 ], 00:29:33.168 "product_name": "Malloc disk", 00:29:33.168 "block_size": 512, 00:29:33.168 "num_blocks": 65536, 00:29:33.168 "uuid": "b79b52eb-a370-4673-b982-de7baf01575b", 00:29:33.168 "assigned_rate_limits": { 00:29:33.168 "rw_ios_per_sec": 0, 00:29:33.168 "rw_mbytes_per_sec": 0, 00:29:33.168 "r_mbytes_per_sec": 0, 00:29:33.168 "w_mbytes_per_sec": 0 00:29:33.168 }, 00:29:33.168 "claimed": true, 00:29:33.168 "claim_type": "exclusive_write", 00:29:33.168 "zoned": false, 00:29:33.168 "supported_io_types": { 00:29:33.168 "read": true, 00:29:33.169 "write": true, 00:29:33.169 "unmap": true, 00:29:33.169 "flush": true, 00:29:33.169 "reset": true, 00:29:33.169 "nvme_admin": false, 00:29:33.169 "nvme_io": false, 00:29:33.169 "nvme_io_md": false, 00:29:33.169 "write_zeroes": true, 00:29:33.169 "zcopy": true, 00:29:33.169 "get_zone_info": false, 00:29:33.169 "zone_management": false, 00:29:33.169 "zone_append": false, 00:29:33.169 "compare": false, 00:29:33.169 "compare_and_write": false, 00:29:33.169 "abort": true, 00:29:33.169 "seek_hole": false, 00:29:33.169 "seek_data": false, 00:29:33.169 "copy": true, 00:29:33.169 "nvme_iov_md": false 00:29:33.169 }, 00:29:33.169 "memory_domains": [ 00:29:33.169 { 00:29:33.169 "dma_device_id": "system", 00:29:33.169 "dma_device_type": 1 00:29:33.169 }, 00:29:33.169 { 00:29:33.169 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:33.169 "dma_device_type": 2 00:29:33.169 } 00:29:33.169 ], 00:29:33.169 "driver_specific": {} 00:29:33.169 } 00:29:33.169 ] 00:29:33.169 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:33.169 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:29:33.169 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:29:33.169 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:29:33.169 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:29:33.169 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:33.169 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:33.169 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:33.169 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:33.169 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:33.169 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:33.169 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:33.169 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:33.169 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:33.169 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:33.169 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:33.169 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:33.169 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:33.169 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:33.169 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:33.169 "name": "Existed_Raid", 00:29:33.169 "uuid": "43b47966-e78f-49ee-980d-85c2383911d1", 00:29:33.169 "strip_size_kb": 64, 00:29:33.169 "state": "online", 00:29:33.169 "raid_level": "concat", 00:29:33.169 "superblock": true, 00:29:33.169 "num_base_bdevs": 4, 00:29:33.169 "num_base_bdevs_discovered": 4, 00:29:33.169 "num_base_bdevs_operational": 4, 00:29:33.169 "base_bdevs_list": [ 00:29:33.169 { 00:29:33.169 "name": "BaseBdev1", 00:29:33.169 "uuid": "7783341b-483f-42da-a0ff-6443837b16bc", 00:29:33.169 "is_configured": true, 00:29:33.169 "data_offset": 2048, 00:29:33.169 "data_size": 63488 00:29:33.169 }, 00:29:33.169 { 00:29:33.169 "name": "BaseBdev2", 00:29:33.169 "uuid": "a6b73762-4a53-4bac-8d7f-66c0efc1e86f", 00:29:33.169 "is_configured": true, 00:29:33.169 "data_offset": 2048, 00:29:33.169 "data_size": 63488 00:29:33.169 }, 00:29:33.169 { 00:29:33.169 "name": "BaseBdev3", 00:29:33.169 "uuid": "b5476039-38ee-4d53-a1a9-d0438b021262", 00:29:33.169 "is_configured": true, 00:29:33.169 "data_offset": 2048, 00:29:33.169 "data_size": 63488 00:29:33.169 }, 00:29:33.169 { 00:29:33.169 "name": "BaseBdev4", 00:29:33.169 "uuid": "b79b52eb-a370-4673-b982-de7baf01575b", 00:29:33.169 "is_configured": true, 00:29:33.169 "data_offset": 2048, 00:29:33.169 "data_size": 63488 00:29:33.169 } 00:29:33.169 ] 00:29:33.169 }' 00:29:33.169 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:33.169 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:33.427 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:29:33.427 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:29:33.427 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:29:33.427 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:29:33.427 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:29:33.427 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:29:33.427 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:29:33.427 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:33.427 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:33.427 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:29:33.427 [2024-11-20 13:49:40.405788] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:33.427 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:33.427 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:33.427 "name": "Existed_Raid", 00:29:33.427 "aliases": [ 00:29:33.427 "43b47966-e78f-49ee-980d-85c2383911d1" 00:29:33.427 ], 00:29:33.427 "product_name": "Raid Volume", 00:29:33.427 "block_size": 512, 00:29:33.427 "num_blocks": 253952, 00:29:33.427 "uuid": "43b47966-e78f-49ee-980d-85c2383911d1", 00:29:33.427 "assigned_rate_limits": { 00:29:33.427 "rw_ios_per_sec": 0, 00:29:33.427 "rw_mbytes_per_sec": 0, 00:29:33.427 "r_mbytes_per_sec": 0, 00:29:33.427 "w_mbytes_per_sec": 0 00:29:33.427 }, 00:29:33.427 "claimed": false, 00:29:33.427 "zoned": false, 00:29:33.427 "supported_io_types": { 00:29:33.427 "read": true, 00:29:33.427 "write": true, 00:29:33.427 "unmap": true, 00:29:33.427 "flush": true, 00:29:33.427 "reset": true, 00:29:33.427 "nvme_admin": false, 00:29:33.427 "nvme_io": false, 00:29:33.427 "nvme_io_md": false, 00:29:33.427 "write_zeroes": true, 00:29:33.427 "zcopy": false, 00:29:33.427 "get_zone_info": false, 00:29:33.427 "zone_management": false, 00:29:33.427 "zone_append": false, 00:29:33.427 "compare": false, 00:29:33.427 "compare_and_write": false, 00:29:33.427 "abort": false, 00:29:33.427 "seek_hole": false, 00:29:33.427 "seek_data": false, 00:29:33.427 "copy": false, 00:29:33.427 "nvme_iov_md": false 00:29:33.427 }, 00:29:33.427 "memory_domains": [ 00:29:33.427 { 00:29:33.427 "dma_device_id": "system", 00:29:33.427 "dma_device_type": 1 00:29:33.427 }, 00:29:33.427 { 00:29:33.427 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:33.427 "dma_device_type": 2 00:29:33.427 }, 00:29:33.427 { 00:29:33.427 "dma_device_id": "system", 00:29:33.427 "dma_device_type": 1 00:29:33.427 }, 00:29:33.427 { 00:29:33.427 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:33.427 "dma_device_type": 2 00:29:33.427 }, 00:29:33.427 { 00:29:33.427 "dma_device_id": "system", 00:29:33.427 "dma_device_type": 1 00:29:33.427 }, 00:29:33.427 { 00:29:33.427 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:33.427 "dma_device_type": 2 00:29:33.427 }, 00:29:33.427 { 00:29:33.427 "dma_device_id": "system", 00:29:33.427 "dma_device_type": 1 00:29:33.427 }, 00:29:33.427 { 00:29:33.427 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:33.427 "dma_device_type": 2 00:29:33.427 } 00:29:33.427 ], 00:29:33.427 "driver_specific": { 00:29:33.427 "raid": { 00:29:33.427 "uuid": "43b47966-e78f-49ee-980d-85c2383911d1", 00:29:33.427 "strip_size_kb": 64, 00:29:33.427 "state": "online", 00:29:33.427 "raid_level": "concat", 00:29:33.427 "superblock": true, 00:29:33.427 "num_base_bdevs": 4, 00:29:33.427 "num_base_bdevs_discovered": 4, 00:29:33.427 "num_base_bdevs_operational": 4, 00:29:33.427 "base_bdevs_list": [ 00:29:33.427 { 00:29:33.427 "name": "BaseBdev1", 00:29:33.427 "uuid": "7783341b-483f-42da-a0ff-6443837b16bc", 00:29:33.427 "is_configured": true, 00:29:33.427 "data_offset": 2048, 00:29:33.427 "data_size": 63488 00:29:33.427 }, 00:29:33.427 { 00:29:33.427 "name": "BaseBdev2", 00:29:33.427 "uuid": "a6b73762-4a53-4bac-8d7f-66c0efc1e86f", 00:29:33.427 "is_configured": true, 00:29:33.428 "data_offset": 2048, 00:29:33.428 "data_size": 63488 00:29:33.428 }, 00:29:33.428 { 00:29:33.428 "name": "BaseBdev3", 00:29:33.428 "uuid": "b5476039-38ee-4d53-a1a9-d0438b021262", 00:29:33.428 "is_configured": true, 00:29:33.428 "data_offset": 2048, 00:29:33.428 "data_size": 63488 00:29:33.428 }, 00:29:33.428 { 00:29:33.428 "name": "BaseBdev4", 00:29:33.428 "uuid": "b79b52eb-a370-4673-b982-de7baf01575b", 00:29:33.428 "is_configured": true, 00:29:33.428 "data_offset": 2048, 00:29:33.428 "data_size": 63488 00:29:33.428 } 00:29:33.428 ] 00:29:33.428 } 00:29:33.428 } 00:29:33.428 }' 00:29:33.428 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:29:33.428 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:29:33.428 BaseBdev2 00:29:33.428 BaseBdev3 00:29:33.428 BaseBdev4' 00:29:33.428 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:33.686 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:33.686 [2024-11-20 13:49:40.637533] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:29:33.687 [2024-11-20 13:49:40.637571] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:29:33.687 [2024-11-20 13:49:40.637627] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:33.687 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:33.687 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:29:33.687 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:29:33.687 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:29:33.687 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:29:33.687 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:29:33.687 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 3 00:29:33.687 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:33.687 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:29:33.687 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:33.687 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:33.687 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:33.687 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:33.687 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:33.687 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:33.687 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:33.687 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:33.687 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:33.687 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:33.687 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:33.687 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:33.687 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:33.687 "name": "Existed_Raid", 00:29:33.687 "uuid": "43b47966-e78f-49ee-980d-85c2383911d1", 00:29:33.687 "strip_size_kb": 64, 00:29:33.687 "state": "offline", 00:29:33.687 "raid_level": "concat", 00:29:33.687 "superblock": true, 00:29:33.687 "num_base_bdevs": 4, 00:29:33.687 "num_base_bdevs_discovered": 3, 00:29:33.687 "num_base_bdevs_operational": 3, 00:29:33.687 "base_bdevs_list": [ 00:29:33.687 { 00:29:33.687 "name": null, 00:29:33.687 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:33.687 "is_configured": false, 00:29:33.687 "data_offset": 0, 00:29:33.687 "data_size": 63488 00:29:33.687 }, 00:29:33.687 { 00:29:33.687 "name": "BaseBdev2", 00:29:33.687 "uuid": "a6b73762-4a53-4bac-8d7f-66c0efc1e86f", 00:29:33.687 "is_configured": true, 00:29:33.687 "data_offset": 2048, 00:29:33.687 "data_size": 63488 00:29:33.687 }, 00:29:33.687 { 00:29:33.687 "name": "BaseBdev3", 00:29:33.687 "uuid": "b5476039-38ee-4d53-a1a9-d0438b021262", 00:29:33.687 "is_configured": true, 00:29:33.687 "data_offset": 2048, 00:29:33.687 "data_size": 63488 00:29:33.687 }, 00:29:33.687 { 00:29:33.687 "name": "BaseBdev4", 00:29:33.687 "uuid": "b79b52eb-a370-4673-b982-de7baf01575b", 00:29:33.687 "is_configured": true, 00:29:33.687 "data_offset": 2048, 00:29:33.687 "data_size": 63488 00:29:33.687 } 00:29:33.687 ] 00:29:33.687 }' 00:29:33.687 13:49:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:33.687 13:49:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:34.254 [2024-11-20 13:49:41.051578] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:34.254 [2024-11-20 13:49:41.156085] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:34.254 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:34.254 [2024-11-20 13:49:41.262041] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:29:34.254 [2024-11-20 13:49:41.262099] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:34.514 BaseBdev2 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:34.514 [ 00:29:34.514 { 00:29:34.514 "name": "BaseBdev2", 00:29:34.514 "aliases": [ 00:29:34.514 "f18bcdae-5514-4258-bac1-372ec6892c60" 00:29:34.514 ], 00:29:34.514 "product_name": "Malloc disk", 00:29:34.514 "block_size": 512, 00:29:34.514 "num_blocks": 65536, 00:29:34.514 "uuid": "f18bcdae-5514-4258-bac1-372ec6892c60", 00:29:34.514 "assigned_rate_limits": { 00:29:34.514 "rw_ios_per_sec": 0, 00:29:34.514 "rw_mbytes_per_sec": 0, 00:29:34.514 "r_mbytes_per_sec": 0, 00:29:34.514 "w_mbytes_per_sec": 0 00:29:34.514 }, 00:29:34.514 "claimed": false, 00:29:34.514 "zoned": false, 00:29:34.514 "supported_io_types": { 00:29:34.514 "read": true, 00:29:34.514 "write": true, 00:29:34.514 "unmap": true, 00:29:34.514 "flush": true, 00:29:34.514 "reset": true, 00:29:34.514 "nvme_admin": false, 00:29:34.514 "nvme_io": false, 00:29:34.514 "nvme_io_md": false, 00:29:34.514 "write_zeroes": true, 00:29:34.514 "zcopy": true, 00:29:34.514 "get_zone_info": false, 00:29:34.514 "zone_management": false, 00:29:34.514 "zone_append": false, 00:29:34.514 "compare": false, 00:29:34.514 "compare_and_write": false, 00:29:34.514 "abort": true, 00:29:34.514 "seek_hole": false, 00:29:34.514 "seek_data": false, 00:29:34.514 "copy": true, 00:29:34.514 "nvme_iov_md": false 00:29:34.514 }, 00:29:34.514 "memory_domains": [ 00:29:34.514 { 00:29:34.514 "dma_device_id": "system", 00:29:34.514 "dma_device_type": 1 00:29:34.514 }, 00:29:34.514 { 00:29:34.514 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:34.514 "dma_device_type": 2 00:29:34.514 } 00:29:34.514 ], 00:29:34.514 "driver_specific": {} 00:29:34.514 } 00:29:34.514 ] 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:34.514 BaseBdev3 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:34.514 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:34.515 [ 00:29:34.515 { 00:29:34.515 "name": "BaseBdev3", 00:29:34.515 "aliases": [ 00:29:34.515 "6a5762f6-f77b-47db-8294-1ed62b1d12f5" 00:29:34.515 ], 00:29:34.515 "product_name": "Malloc disk", 00:29:34.515 "block_size": 512, 00:29:34.515 "num_blocks": 65536, 00:29:34.515 "uuid": "6a5762f6-f77b-47db-8294-1ed62b1d12f5", 00:29:34.515 "assigned_rate_limits": { 00:29:34.515 "rw_ios_per_sec": 0, 00:29:34.515 "rw_mbytes_per_sec": 0, 00:29:34.515 "r_mbytes_per_sec": 0, 00:29:34.515 "w_mbytes_per_sec": 0 00:29:34.515 }, 00:29:34.515 "claimed": false, 00:29:34.515 "zoned": false, 00:29:34.515 "supported_io_types": { 00:29:34.515 "read": true, 00:29:34.515 "write": true, 00:29:34.515 "unmap": true, 00:29:34.515 "flush": true, 00:29:34.515 "reset": true, 00:29:34.515 "nvme_admin": false, 00:29:34.515 "nvme_io": false, 00:29:34.515 "nvme_io_md": false, 00:29:34.515 "write_zeroes": true, 00:29:34.515 "zcopy": true, 00:29:34.515 "get_zone_info": false, 00:29:34.515 "zone_management": false, 00:29:34.515 "zone_append": false, 00:29:34.515 "compare": false, 00:29:34.515 "compare_and_write": false, 00:29:34.515 "abort": true, 00:29:34.515 "seek_hole": false, 00:29:34.515 "seek_data": false, 00:29:34.515 "copy": true, 00:29:34.515 "nvme_iov_md": false 00:29:34.515 }, 00:29:34.515 "memory_domains": [ 00:29:34.515 { 00:29:34.515 "dma_device_id": "system", 00:29:34.515 "dma_device_type": 1 00:29:34.515 }, 00:29:34.515 { 00:29:34.515 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:34.515 "dma_device_type": 2 00:29:34.515 } 00:29:34.515 ], 00:29:34.515 "driver_specific": {} 00:29:34.515 } 00:29:34.515 ] 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:34.515 BaseBdev4 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:34.515 [ 00:29:34.515 { 00:29:34.515 "name": "BaseBdev4", 00:29:34.515 "aliases": [ 00:29:34.515 "1feedd38-7e3f-4f4e-9fba-a20fbd02d07a" 00:29:34.515 ], 00:29:34.515 "product_name": "Malloc disk", 00:29:34.515 "block_size": 512, 00:29:34.515 "num_blocks": 65536, 00:29:34.515 "uuid": "1feedd38-7e3f-4f4e-9fba-a20fbd02d07a", 00:29:34.515 "assigned_rate_limits": { 00:29:34.515 "rw_ios_per_sec": 0, 00:29:34.515 "rw_mbytes_per_sec": 0, 00:29:34.515 "r_mbytes_per_sec": 0, 00:29:34.515 "w_mbytes_per_sec": 0 00:29:34.515 }, 00:29:34.515 "claimed": false, 00:29:34.515 "zoned": false, 00:29:34.515 "supported_io_types": { 00:29:34.515 "read": true, 00:29:34.515 "write": true, 00:29:34.515 "unmap": true, 00:29:34.515 "flush": true, 00:29:34.515 "reset": true, 00:29:34.515 "nvme_admin": false, 00:29:34.515 "nvme_io": false, 00:29:34.515 "nvme_io_md": false, 00:29:34.515 "write_zeroes": true, 00:29:34.515 "zcopy": true, 00:29:34.515 "get_zone_info": false, 00:29:34.515 "zone_management": false, 00:29:34.515 "zone_append": false, 00:29:34.515 "compare": false, 00:29:34.515 "compare_and_write": false, 00:29:34.515 "abort": true, 00:29:34.515 "seek_hole": false, 00:29:34.515 "seek_data": false, 00:29:34.515 "copy": true, 00:29:34.515 "nvme_iov_md": false 00:29:34.515 }, 00:29:34.515 "memory_domains": [ 00:29:34.515 { 00:29:34.515 "dma_device_id": "system", 00:29:34.515 "dma_device_type": 1 00:29:34.515 }, 00:29:34.515 { 00:29:34.515 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:34.515 "dma_device_type": 2 00:29:34.515 } 00:29:34.515 ], 00:29:34.515 "driver_specific": {} 00:29:34.515 } 00:29:34.515 ] 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:34.515 [2024-11-20 13:49:41.528035] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:29:34.515 [2024-11-20 13:49:41.528089] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:29:34.515 [2024-11-20 13:49:41.528113] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:29:34.515 [2024-11-20 13:49:41.530112] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:29:34.515 [2024-11-20 13:49:41.530170] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:34.515 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:34.515 "name": "Existed_Raid", 00:29:34.515 "uuid": "28351bdb-a74c-4266-81de-8a46f9d420e0", 00:29:34.515 "strip_size_kb": 64, 00:29:34.515 "state": "configuring", 00:29:34.515 "raid_level": "concat", 00:29:34.515 "superblock": true, 00:29:34.515 "num_base_bdevs": 4, 00:29:34.515 "num_base_bdevs_discovered": 3, 00:29:34.515 "num_base_bdevs_operational": 4, 00:29:34.515 "base_bdevs_list": [ 00:29:34.515 { 00:29:34.515 "name": "BaseBdev1", 00:29:34.515 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:34.515 "is_configured": false, 00:29:34.515 "data_offset": 0, 00:29:34.515 "data_size": 0 00:29:34.515 }, 00:29:34.515 { 00:29:34.515 "name": "BaseBdev2", 00:29:34.515 "uuid": "f18bcdae-5514-4258-bac1-372ec6892c60", 00:29:34.515 "is_configured": true, 00:29:34.515 "data_offset": 2048, 00:29:34.515 "data_size": 63488 00:29:34.515 }, 00:29:34.515 { 00:29:34.515 "name": "BaseBdev3", 00:29:34.515 "uuid": "6a5762f6-f77b-47db-8294-1ed62b1d12f5", 00:29:34.515 "is_configured": true, 00:29:34.515 "data_offset": 2048, 00:29:34.515 "data_size": 63488 00:29:34.515 }, 00:29:34.515 { 00:29:34.515 "name": "BaseBdev4", 00:29:34.515 "uuid": "1feedd38-7e3f-4f4e-9fba-a20fbd02d07a", 00:29:34.516 "is_configured": true, 00:29:34.516 "data_offset": 2048, 00:29:34.516 "data_size": 63488 00:29:34.516 } 00:29:34.516 ] 00:29:34.516 }' 00:29:34.516 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:34.516 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:35.081 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:29:35.081 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:35.081 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:35.081 [2024-11-20 13:49:41.840128] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:29:35.081 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:35.081 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:29:35.081 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:35.081 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:35.081 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:35.081 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:35.081 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:35.081 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:35.081 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:35.081 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:35.081 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:35.081 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:35.081 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:35.081 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:35.081 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:35.081 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:35.081 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:35.081 "name": "Existed_Raid", 00:29:35.081 "uuid": "28351bdb-a74c-4266-81de-8a46f9d420e0", 00:29:35.081 "strip_size_kb": 64, 00:29:35.081 "state": "configuring", 00:29:35.081 "raid_level": "concat", 00:29:35.081 "superblock": true, 00:29:35.081 "num_base_bdevs": 4, 00:29:35.081 "num_base_bdevs_discovered": 2, 00:29:35.081 "num_base_bdevs_operational": 4, 00:29:35.081 "base_bdevs_list": [ 00:29:35.081 { 00:29:35.081 "name": "BaseBdev1", 00:29:35.081 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:35.081 "is_configured": false, 00:29:35.081 "data_offset": 0, 00:29:35.081 "data_size": 0 00:29:35.081 }, 00:29:35.081 { 00:29:35.081 "name": null, 00:29:35.081 "uuid": "f18bcdae-5514-4258-bac1-372ec6892c60", 00:29:35.081 "is_configured": false, 00:29:35.081 "data_offset": 0, 00:29:35.081 "data_size": 63488 00:29:35.081 }, 00:29:35.081 { 00:29:35.081 "name": "BaseBdev3", 00:29:35.081 "uuid": "6a5762f6-f77b-47db-8294-1ed62b1d12f5", 00:29:35.081 "is_configured": true, 00:29:35.081 "data_offset": 2048, 00:29:35.081 "data_size": 63488 00:29:35.081 }, 00:29:35.081 { 00:29:35.081 "name": "BaseBdev4", 00:29:35.081 "uuid": "1feedd38-7e3f-4f4e-9fba-a20fbd02d07a", 00:29:35.081 "is_configured": true, 00:29:35.081 "data_offset": 2048, 00:29:35.081 "data_size": 63488 00:29:35.081 } 00:29:35.081 ] 00:29:35.081 }' 00:29:35.081 13:49:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:35.081 13:49:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:35.339 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:35.339 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:29:35.339 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:35.339 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:35.339 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:35.339 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:29:35.339 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:29:35.339 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:35.339 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:35.339 [2024-11-20 13:49:42.217287] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:29:35.339 BaseBdev1 00:29:35.339 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:35.339 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:29:35.339 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:29:35.339 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:35.339 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:29:35.339 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:35.339 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:35.339 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:35.339 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:35.339 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:35.339 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:35.339 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:29:35.339 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:35.339 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:35.339 [ 00:29:35.339 { 00:29:35.339 "name": "BaseBdev1", 00:29:35.339 "aliases": [ 00:29:35.339 "777ef82d-ee3a-4ba4-926c-c6d81d3ade5b" 00:29:35.340 ], 00:29:35.340 "product_name": "Malloc disk", 00:29:35.340 "block_size": 512, 00:29:35.340 "num_blocks": 65536, 00:29:35.340 "uuid": "777ef82d-ee3a-4ba4-926c-c6d81d3ade5b", 00:29:35.340 "assigned_rate_limits": { 00:29:35.340 "rw_ios_per_sec": 0, 00:29:35.340 "rw_mbytes_per_sec": 0, 00:29:35.340 "r_mbytes_per_sec": 0, 00:29:35.340 "w_mbytes_per_sec": 0 00:29:35.340 }, 00:29:35.340 "claimed": true, 00:29:35.340 "claim_type": "exclusive_write", 00:29:35.340 "zoned": false, 00:29:35.340 "supported_io_types": { 00:29:35.340 "read": true, 00:29:35.340 "write": true, 00:29:35.340 "unmap": true, 00:29:35.340 "flush": true, 00:29:35.340 "reset": true, 00:29:35.340 "nvme_admin": false, 00:29:35.340 "nvme_io": false, 00:29:35.340 "nvme_io_md": false, 00:29:35.340 "write_zeroes": true, 00:29:35.340 "zcopy": true, 00:29:35.340 "get_zone_info": false, 00:29:35.340 "zone_management": false, 00:29:35.340 "zone_append": false, 00:29:35.340 "compare": false, 00:29:35.340 "compare_and_write": false, 00:29:35.340 "abort": true, 00:29:35.340 "seek_hole": false, 00:29:35.340 "seek_data": false, 00:29:35.340 "copy": true, 00:29:35.340 "nvme_iov_md": false 00:29:35.340 }, 00:29:35.340 "memory_domains": [ 00:29:35.340 { 00:29:35.340 "dma_device_id": "system", 00:29:35.340 "dma_device_type": 1 00:29:35.340 }, 00:29:35.340 { 00:29:35.340 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:35.340 "dma_device_type": 2 00:29:35.340 } 00:29:35.340 ], 00:29:35.340 "driver_specific": {} 00:29:35.340 } 00:29:35.340 ] 00:29:35.340 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:35.340 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:29:35.340 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:29:35.340 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:35.340 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:35.340 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:35.340 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:35.340 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:35.340 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:35.340 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:35.340 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:35.340 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:35.340 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:35.340 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:35.340 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:35.340 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:35.340 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:35.340 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:35.340 "name": "Existed_Raid", 00:29:35.340 "uuid": "28351bdb-a74c-4266-81de-8a46f9d420e0", 00:29:35.340 "strip_size_kb": 64, 00:29:35.340 "state": "configuring", 00:29:35.340 "raid_level": "concat", 00:29:35.340 "superblock": true, 00:29:35.340 "num_base_bdevs": 4, 00:29:35.340 "num_base_bdevs_discovered": 3, 00:29:35.340 "num_base_bdevs_operational": 4, 00:29:35.340 "base_bdevs_list": [ 00:29:35.340 { 00:29:35.340 "name": "BaseBdev1", 00:29:35.340 "uuid": "777ef82d-ee3a-4ba4-926c-c6d81d3ade5b", 00:29:35.340 "is_configured": true, 00:29:35.340 "data_offset": 2048, 00:29:35.340 "data_size": 63488 00:29:35.340 }, 00:29:35.340 { 00:29:35.340 "name": null, 00:29:35.340 "uuid": "f18bcdae-5514-4258-bac1-372ec6892c60", 00:29:35.340 "is_configured": false, 00:29:35.340 "data_offset": 0, 00:29:35.340 "data_size": 63488 00:29:35.340 }, 00:29:35.340 { 00:29:35.340 "name": "BaseBdev3", 00:29:35.340 "uuid": "6a5762f6-f77b-47db-8294-1ed62b1d12f5", 00:29:35.340 "is_configured": true, 00:29:35.340 "data_offset": 2048, 00:29:35.340 "data_size": 63488 00:29:35.340 }, 00:29:35.340 { 00:29:35.340 "name": "BaseBdev4", 00:29:35.340 "uuid": "1feedd38-7e3f-4f4e-9fba-a20fbd02d07a", 00:29:35.340 "is_configured": true, 00:29:35.340 "data_offset": 2048, 00:29:35.340 "data_size": 63488 00:29:35.340 } 00:29:35.340 ] 00:29:35.340 }' 00:29:35.340 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:35.340 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:35.598 [2024-11-20 13:49:42.593510] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:35.598 "name": "Existed_Raid", 00:29:35.598 "uuid": "28351bdb-a74c-4266-81de-8a46f9d420e0", 00:29:35.598 "strip_size_kb": 64, 00:29:35.598 "state": "configuring", 00:29:35.598 "raid_level": "concat", 00:29:35.598 "superblock": true, 00:29:35.598 "num_base_bdevs": 4, 00:29:35.598 "num_base_bdevs_discovered": 2, 00:29:35.598 "num_base_bdevs_operational": 4, 00:29:35.598 "base_bdevs_list": [ 00:29:35.598 { 00:29:35.598 "name": "BaseBdev1", 00:29:35.598 "uuid": "777ef82d-ee3a-4ba4-926c-c6d81d3ade5b", 00:29:35.598 "is_configured": true, 00:29:35.598 "data_offset": 2048, 00:29:35.598 "data_size": 63488 00:29:35.598 }, 00:29:35.598 { 00:29:35.598 "name": null, 00:29:35.598 "uuid": "f18bcdae-5514-4258-bac1-372ec6892c60", 00:29:35.598 "is_configured": false, 00:29:35.598 "data_offset": 0, 00:29:35.598 "data_size": 63488 00:29:35.598 }, 00:29:35.598 { 00:29:35.598 "name": null, 00:29:35.598 "uuid": "6a5762f6-f77b-47db-8294-1ed62b1d12f5", 00:29:35.598 "is_configured": false, 00:29:35.598 "data_offset": 0, 00:29:35.598 "data_size": 63488 00:29:35.598 }, 00:29:35.598 { 00:29:35.598 "name": "BaseBdev4", 00:29:35.598 "uuid": "1feedd38-7e3f-4f4e-9fba-a20fbd02d07a", 00:29:35.598 "is_configured": true, 00:29:35.598 "data_offset": 2048, 00:29:35.598 "data_size": 63488 00:29:35.598 } 00:29:35.598 ] 00:29:35.598 }' 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:35.598 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:35.856 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:35.856 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:35.856 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:35.856 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:29:35.857 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:36.114 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:29:36.114 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:29:36.114 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:36.114 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:36.115 [2024-11-20 13:49:42.929559] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:29:36.115 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:36.115 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:29:36.115 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:36.115 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:36.115 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:36.115 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:36.115 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:36.115 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:36.115 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:36.115 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:36.115 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:36.115 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:36.115 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:36.115 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:36.115 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:36.115 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:36.115 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:36.115 "name": "Existed_Raid", 00:29:36.115 "uuid": "28351bdb-a74c-4266-81de-8a46f9d420e0", 00:29:36.115 "strip_size_kb": 64, 00:29:36.115 "state": "configuring", 00:29:36.115 "raid_level": "concat", 00:29:36.115 "superblock": true, 00:29:36.115 "num_base_bdevs": 4, 00:29:36.115 "num_base_bdevs_discovered": 3, 00:29:36.115 "num_base_bdevs_operational": 4, 00:29:36.115 "base_bdevs_list": [ 00:29:36.115 { 00:29:36.115 "name": "BaseBdev1", 00:29:36.115 "uuid": "777ef82d-ee3a-4ba4-926c-c6d81d3ade5b", 00:29:36.115 "is_configured": true, 00:29:36.115 "data_offset": 2048, 00:29:36.115 "data_size": 63488 00:29:36.115 }, 00:29:36.115 { 00:29:36.115 "name": null, 00:29:36.115 "uuid": "f18bcdae-5514-4258-bac1-372ec6892c60", 00:29:36.115 "is_configured": false, 00:29:36.115 "data_offset": 0, 00:29:36.115 "data_size": 63488 00:29:36.115 }, 00:29:36.115 { 00:29:36.115 "name": "BaseBdev3", 00:29:36.115 "uuid": "6a5762f6-f77b-47db-8294-1ed62b1d12f5", 00:29:36.115 "is_configured": true, 00:29:36.115 "data_offset": 2048, 00:29:36.115 "data_size": 63488 00:29:36.115 }, 00:29:36.115 { 00:29:36.115 "name": "BaseBdev4", 00:29:36.115 "uuid": "1feedd38-7e3f-4f4e-9fba-a20fbd02d07a", 00:29:36.115 "is_configured": true, 00:29:36.115 "data_offset": 2048, 00:29:36.115 "data_size": 63488 00:29:36.115 } 00:29:36.115 ] 00:29:36.115 }' 00:29:36.115 13:49:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:36.115 13:49:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:36.373 [2024-11-20 13:49:43.281656] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:36.373 "name": "Existed_Raid", 00:29:36.373 "uuid": "28351bdb-a74c-4266-81de-8a46f9d420e0", 00:29:36.373 "strip_size_kb": 64, 00:29:36.373 "state": "configuring", 00:29:36.373 "raid_level": "concat", 00:29:36.373 "superblock": true, 00:29:36.373 "num_base_bdevs": 4, 00:29:36.373 "num_base_bdevs_discovered": 2, 00:29:36.373 "num_base_bdevs_operational": 4, 00:29:36.373 "base_bdevs_list": [ 00:29:36.373 { 00:29:36.373 "name": null, 00:29:36.373 "uuid": "777ef82d-ee3a-4ba4-926c-c6d81d3ade5b", 00:29:36.373 "is_configured": false, 00:29:36.373 "data_offset": 0, 00:29:36.373 "data_size": 63488 00:29:36.373 }, 00:29:36.373 { 00:29:36.373 "name": null, 00:29:36.373 "uuid": "f18bcdae-5514-4258-bac1-372ec6892c60", 00:29:36.373 "is_configured": false, 00:29:36.373 "data_offset": 0, 00:29:36.373 "data_size": 63488 00:29:36.373 }, 00:29:36.373 { 00:29:36.373 "name": "BaseBdev3", 00:29:36.373 "uuid": "6a5762f6-f77b-47db-8294-1ed62b1d12f5", 00:29:36.373 "is_configured": true, 00:29:36.373 "data_offset": 2048, 00:29:36.373 "data_size": 63488 00:29:36.373 }, 00:29:36.373 { 00:29:36.373 "name": "BaseBdev4", 00:29:36.373 "uuid": "1feedd38-7e3f-4f4e-9fba-a20fbd02d07a", 00:29:36.373 "is_configured": true, 00:29:36.373 "data_offset": 2048, 00:29:36.373 "data_size": 63488 00:29:36.373 } 00:29:36.373 ] 00:29:36.373 }' 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:36.373 13:49:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:36.631 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:36.631 13:49:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:36.631 13:49:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:36.631 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:29:36.631 13:49:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:36.631 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:29:36.631 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:29:36.631 13:49:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:36.631 13:49:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:36.631 [2024-11-20 13:49:43.684646] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:29:36.889 13:49:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:36.889 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:29:36.889 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:36.889 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:36.889 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:36.889 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:36.889 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:36.889 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:36.889 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:36.889 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:36.889 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:36.889 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:36.889 13:49:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:36.889 13:49:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:36.889 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:36.889 13:49:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:36.889 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:36.889 "name": "Existed_Raid", 00:29:36.889 "uuid": "28351bdb-a74c-4266-81de-8a46f9d420e0", 00:29:36.889 "strip_size_kb": 64, 00:29:36.889 "state": "configuring", 00:29:36.889 "raid_level": "concat", 00:29:36.889 "superblock": true, 00:29:36.889 "num_base_bdevs": 4, 00:29:36.889 "num_base_bdevs_discovered": 3, 00:29:36.889 "num_base_bdevs_operational": 4, 00:29:36.889 "base_bdevs_list": [ 00:29:36.889 { 00:29:36.889 "name": null, 00:29:36.889 "uuid": "777ef82d-ee3a-4ba4-926c-c6d81d3ade5b", 00:29:36.889 "is_configured": false, 00:29:36.889 "data_offset": 0, 00:29:36.889 "data_size": 63488 00:29:36.889 }, 00:29:36.889 { 00:29:36.889 "name": "BaseBdev2", 00:29:36.889 "uuid": "f18bcdae-5514-4258-bac1-372ec6892c60", 00:29:36.889 "is_configured": true, 00:29:36.889 "data_offset": 2048, 00:29:36.889 "data_size": 63488 00:29:36.889 }, 00:29:36.889 { 00:29:36.889 "name": "BaseBdev3", 00:29:36.889 "uuid": "6a5762f6-f77b-47db-8294-1ed62b1d12f5", 00:29:36.889 "is_configured": true, 00:29:36.889 "data_offset": 2048, 00:29:36.889 "data_size": 63488 00:29:36.889 }, 00:29:36.889 { 00:29:36.889 "name": "BaseBdev4", 00:29:36.889 "uuid": "1feedd38-7e3f-4f4e-9fba-a20fbd02d07a", 00:29:36.889 "is_configured": true, 00:29:36.889 "data_offset": 2048, 00:29:36.889 "data_size": 63488 00:29:36.889 } 00:29:36.889 ] 00:29:36.889 }' 00:29:36.889 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:36.889 13:49:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:37.148 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:37.148 13:49:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:29:37.148 13:49:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:37.148 13:49:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 777ef82d-ee3a-4ba4-926c-c6d81d3ade5b 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:37.148 [2024-11-20 13:49:44.085445] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:29:37.148 [2024-11-20 13:49:44.085655] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:29:37.148 [2024-11-20 13:49:44.085666] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:29:37.148 NewBaseBdev 00:29:37.148 [2024-11-20 13:49:44.085901] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:29:37.148 [2024-11-20 13:49:44.086035] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:29:37.148 [2024-11-20 13:49:44.086052] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:29:37.148 [2024-11-20 13:49:44.086159] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:37.148 [ 00:29:37.148 { 00:29:37.148 "name": "NewBaseBdev", 00:29:37.148 "aliases": [ 00:29:37.148 "777ef82d-ee3a-4ba4-926c-c6d81d3ade5b" 00:29:37.148 ], 00:29:37.148 "product_name": "Malloc disk", 00:29:37.148 "block_size": 512, 00:29:37.148 "num_blocks": 65536, 00:29:37.148 "uuid": "777ef82d-ee3a-4ba4-926c-c6d81d3ade5b", 00:29:37.148 "assigned_rate_limits": { 00:29:37.148 "rw_ios_per_sec": 0, 00:29:37.148 "rw_mbytes_per_sec": 0, 00:29:37.148 "r_mbytes_per_sec": 0, 00:29:37.148 "w_mbytes_per_sec": 0 00:29:37.148 }, 00:29:37.148 "claimed": true, 00:29:37.148 "claim_type": "exclusive_write", 00:29:37.148 "zoned": false, 00:29:37.148 "supported_io_types": { 00:29:37.148 "read": true, 00:29:37.148 "write": true, 00:29:37.148 "unmap": true, 00:29:37.148 "flush": true, 00:29:37.148 "reset": true, 00:29:37.148 "nvme_admin": false, 00:29:37.148 "nvme_io": false, 00:29:37.148 "nvme_io_md": false, 00:29:37.148 "write_zeroes": true, 00:29:37.148 "zcopy": true, 00:29:37.148 "get_zone_info": false, 00:29:37.148 "zone_management": false, 00:29:37.148 "zone_append": false, 00:29:37.148 "compare": false, 00:29:37.148 "compare_and_write": false, 00:29:37.148 "abort": true, 00:29:37.148 "seek_hole": false, 00:29:37.148 "seek_data": false, 00:29:37.148 "copy": true, 00:29:37.148 "nvme_iov_md": false 00:29:37.148 }, 00:29:37.148 "memory_domains": [ 00:29:37.148 { 00:29:37.148 "dma_device_id": "system", 00:29:37.148 "dma_device_type": 1 00:29:37.148 }, 00:29:37.148 { 00:29:37.148 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:37.148 "dma_device_type": 2 00:29:37.148 } 00:29:37.148 ], 00:29:37.148 "driver_specific": {} 00:29:37.148 } 00:29:37.148 ] 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:37.148 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:37.149 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:37.149 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:37.149 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:37.149 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:37.149 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:37.149 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:37.149 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:37.149 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:37.149 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:37.149 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:37.149 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:37.149 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:37.149 "name": "Existed_Raid", 00:29:37.149 "uuid": "28351bdb-a74c-4266-81de-8a46f9d420e0", 00:29:37.149 "strip_size_kb": 64, 00:29:37.149 "state": "online", 00:29:37.149 "raid_level": "concat", 00:29:37.149 "superblock": true, 00:29:37.149 "num_base_bdevs": 4, 00:29:37.149 "num_base_bdevs_discovered": 4, 00:29:37.149 "num_base_bdevs_operational": 4, 00:29:37.149 "base_bdevs_list": [ 00:29:37.149 { 00:29:37.149 "name": "NewBaseBdev", 00:29:37.149 "uuid": "777ef82d-ee3a-4ba4-926c-c6d81d3ade5b", 00:29:37.149 "is_configured": true, 00:29:37.149 "data_offset": 2048, 00:29:37.149 "data_size": 63488 00:29:37.149 }, 00:29:37.149 { 00:29:37.149 "name": "BaseBdev2", 00:29:37.149 "uuid": "f18bcdae-5514-4258-bac1-372ec6892c60", 00:29:37.149 "is_configured": true, 00:29:37.149 "data_offset": 2048, 00:29:37.149 "data_size": 63488 00:29:37.149 }, 00:29:37.149 { 00:29:37.149 "name": "BaseBdev3", 00:29:37.149 "uuid": "6a5762f6-f77b-47db-8294-1ed62b1d12f5", 00:29:37.149 "is_configured": true, 00:29:37.149 "data_offset": 2048, 00:29:37.149 "data_size": 63488 00:29:37.149 }, 00:29:37.149 { 00:29:37.149 "name": "BaseBdev4", 00:29:37.149 "uuid": "1feedd38-7e3f-4f4e-9fba-a20fbd02d07a", 00:29:37.149 "is_configured": true, 00:29:37.149 "data_offset": 2048, 00:29:37.149 "data_size": 63488 00:29:37.149 } 00:29:37.149 ] 00:29:37.149 }' 00:29:37.149 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:37.149 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:37.407 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:29:37.407 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:29:37.407 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:29:37.407 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:29:37.407 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:29:37.407 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:29:37.407 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:29:37.407 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:37.407 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:37.407 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:29:37.407 [2024-11-20 13:49:44.425908] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:37.407 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:37.407 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:37.407 "name": "Existed_Raid", 00:29:37.407 "aliases": [ 00:29:37.407 "28351bdb-a74c-4266-81de-8a46f9d420e0" 00:29:37.407 ], 00:29:37.407 "product_name": "Raid Volume", 00:29:37.407 "block_size": 512, 00:29:37.407 "num_blocks": 253952, 00:29:37.407 "uuid": "28351bdb-a74c-4266-81de-8a46f9d420e0", 00:29:37.407 "assigned_rate_limits": { 00:29:37.407 "rw_ios_per_sec": 0, 00:29:37.407 "rw_mbytes_per_sec": 0, 00:29:37.407 "r_mbytes_per_sec": 0, 00:29:37.407 "w_mbytes_per_sec": 0 00:29:37.407 }, 00:29:37.407 "claimed": false, 00:29:37.407 "zoned": false, 00:29:37.407 "supported_io_types": { 00:29:37.407 "read": true, 00:29:37.407 "write": true, 00:29:37.407 "unmap": true, 00:29:37.407 "flush": true, 00:29:37.407 "reset": true, 00:29:37.407 "nvme_admin": false, 00:29:37.407 "nvme_io": false, 00:29:37.407 "nvme_io_md": false, 00:29:37.407 "write_zeroes": true, 00:29:37.407 "zcopy": false, 00:29:37.407 "get_zone_info": false, 00:29:37.407 "zone_management": false, 00:29:37.407 "zone_append": false, 00:29:37.407 "compare": false, 00:29:37.407 "compare_and_write": false, 00:29:37.407 "abort": false, 00:29:37.407 "seek_hole": false, 00:29:37.407 "seek_data": false, 00:29:37.407 "copy": false, 00:29:37.407 "nvme_iov_md": false 00:29:37.407 }, 00:29:37.407 "memory_domains": [ 00:29:37.407 { 00:29:37.407 "dma_device_id": "system", 00:29:37.407 "dma_device_type": 1 00:29:37.407 }, 00:29:37.407 { 00:29:37.407 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:37.407 "dma_device_type": 2 00:29:37.407 }, 00:29:37.407 { 00:29:37.407 "dma_device_id": "system", 00:29:37.407 "dma_device_type": 1 00:29:37.407 }, 00:29:37.407 { 00:29:37.407 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:37.407 "dma_device_type": 2 00:29:37.407 }, 00:29:37.407 { 00:29:37.407 "dma_device_id": "system", 00:29:37.407 "dma_device_type": 1 00:29:37.407 }, 00:29:37.407 { 00:29:37.407 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:37.407 "dma_device_type": 2 00:29:37.407 }, 00:29:37.407 { 00:29:37.407 "dma_device_id": "system", 00:29:37.407 "dma_device_type": 1 00:29:37.407 }, 00:29:37.407 { 00:29:37.407 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:37.407 "dma_device_type": 2 00:29:37.407 } 00:29:37.407 ], 00:29:37.407 "driver_specific": { 00:29:37.407 "raid": { 00:29:37.407 "uuid": "28351bdb-a74c-4266-81de-8a46f9d420e0", 00:29:37.407 "strip_size_kb": 64, 00:29:37.407 "state": "online", 00:29:37.407 "raid_level": "concat", 00:29:37.407 "superblock": true, 00:29:37.407 "num_base_bdevs": 4, 00:29:37.407 "num_base_bdevs_discovered": 4, 00:29:37.407 "num_base_bdevs_operational": 4, 00:29:37.407 "base_bdevs_list": [ 00:29:37.407 { 00:29:37.407 "name": "NewBaseBdev", 00:29:37.407 "uuid": "777ef82d-ee3a-4ba4-926c-c6d81d3ade5b", 00:29:37.407 "is_configured": true, 00:29:37.407 "data_offset": 2048, 00:29:37.407 "data_size": 63488 00:29:37.407 }, 00:29:37.407 { 00:29:37.407 "name": "BaseBdev2", 00:29:37.407 "uuid": "f18bcdae-5514-4258-bac1-372ec6892c60", 00:29:37.407 "is_configured": true, 00:29:37.407 "data_offset": 2048, 00:29:37.407 "data_size": 63488 00:29:37.407 }, 00:29:37.407 { 00:29:37.407 "name": "BaseBdev3", 00:29:37.407 "uuid": "6a5762f6-f77b-47db-8294-1ed62b1d12f5", 00:29:37.407 "is_configured": true, 00:29:37.407 "data_offset": 2048, 00:29:37.407 "data_size": 63488 00:29:37.407 }, 00:29:37.407 { 00:29:37.407 "name": "BaseBdev4", 00:29:37.407 "uuid": "1feedd38-7e3f-4f4e-9fba-a20fbd02d07a", 00:29:37.407 "is_configured": true, 00:29:37.407 "data_offset": 2048, 00:29:37.407 "data_size": 63488 00:29:37.407 } 00:29:37.407 ] 00:29:37.407 } 00:29:37.407 } 00:29:37.407 }' 00:29:37.407 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:29:37.665 BaseBdev2 00:29:37.665 BaseBdev3 00:29:37.665 BaseBdev4' 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:37.665 [2024-11-20 13:49:44.665617] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:29:37.665 [2024-11-20 13:49:44.665649] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:29:37.665 [2024-11-20 13:49:44.665734] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:37.665 [2024-11-20 13:49:44.665808] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:29:37.665 [2024-11-20 13:49:44.665818] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 70049 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 70049 ']' 00:29:37.665 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 70049 00:29:37.666 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:29:37.666 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:29:37.666 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70049 00:29:37.666 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:29:37.666 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:29:37.666 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70049' 00:29:37.666 killing process with pid 70049 00:29:37.666 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 70049 00:29:37.666 [2024-11-20 13:49:44.693619] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:29:37.666 13:49:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 70049 00:29:37.923 [2024-11-20 13:49:44.909903] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:29:38.488 13:49:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:29:38.488 00:29:38.488 real 0m8.222s 00:29:38.488 user 0m13.015s 00:29:38.488 sys 0m1.437s 00:29:38.488 13:49:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:29:38.488 ************************************ 00:29:38.488 END TEST raid_state_function_test_sb 00:29:38.488 ************************************ 00:29:38.488 13:49:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:38.747 13:49:45 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 4 00:29:38.747 13:49:45 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:29:38.747 13:49:45 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:29:38.747 13:49:45 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:29:38.747 ************************************ 00:29:38.747 START TEST raid_superblock_test 00:29:38.747 ************************************ 00:29:38.747 13:49:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test concat 4 00:29:38.747 13:49:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:29:38.747 13:49:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:29:38.747 13:49:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:29:38.747 13:49:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:29:38.747 13:49:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:29:38.747 13:49:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:29:38.747 13:49:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:29:38.747 13:49:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:29:38.747 13:49:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:29:38.747 13:49:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:29:38.747 13:49:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:29:38.747 13:49:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:29:38.747 13:49:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:29:38.747 13:49:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:29:38.747 13:49:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:29:38.747 13:49:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:29:38.747 13:49:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=70681 00:29:38.747 13:49:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 70681 00:29:38.747 13:49:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 70681 ']' 00:29:38.747 13:49:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:38.747 13:49:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:29:38.747 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:38.747 13:49:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:29:38.747 13:49:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:38.747 13:49:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:29:38.747 13:49:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:38.747 [2024-11-20 13:49:45.653915] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:29:38.747 [2024-11-20 13:49:45.654051] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70681 ] 00:29:39.032 [2024-11-20 13:49:45.805297] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:39.032 [2024-11-20 13:49:45.913851] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:29:39.032 [2024-11-20 13:49:46.048715] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:39.032 [2024-11-20 13:49:46.048785] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:39.601 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:29:39.601 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:29:39.601 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:29:39.601 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:39.602 malloc1 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:39.602 [2024-11-20 13:49:46.497356] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:29:39.602 [2024-11-20 13:49:46.497429] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:39.602 [2024-11-20 13:49:46.497451] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:29:39.602 [2024-11-20 13:49:46.497459] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:39.602 [2024-11-20 13:49:46.499441] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:39.602 [2024-11-20 13:49:46.499476] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:29:39.602 pt1 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:39.602 malloc2 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:39.602 [2024-11-20 13:49:46.531669] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:29:39.602 [2024-11-20 13:49:46.531727] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:39.602 [2024-11-20 13:49:46.531750] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:29:39.602 [2024-11-20 13:49:46.531758] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:39.602 [2024-11-20 13:49:46.533701] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:39.602 [2024-11-20 13:49:46.533732] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:29:39.602 pt2 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:39.602 malloc3 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:39.602 [2024-11-20 13:49:46.583102] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:29:39.602 [2024-11-20 13:49:46.583175] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:39.602 [2024-11-20 13:49:46.583199] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:29:39.602 [2024-11-20 13:49:46.583207] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:39.602 [2024-11-20 13:49:46.585238] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:39.602 [2024-11-20 13:49:46.585271] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:29:39.602 pt3 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:39.602 malloc4 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:39.602 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:39.602 [2024-11-20 13:49:46.617230] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:29:39.602 [2024-11-20 13:49:46.617288] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:39.602 [2024-11-20 13:49:46.617305] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:29:39.602 [2024-11-20 13:49:46.617313] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:39.603 [2024-11-20 13:49:46.619232] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:39.603 [2024-11-20 13:49:46.619262] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:29:39.603 pt4 00:29:39.603 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:39.603 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:29:39.603 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:29:39.603 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:29:39.603 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:39.603 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:39.603 [2024-11-20 13:49:46.625257] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:29:39.603 [2024-11-20 13:49:46.626884] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:29:39.603 [2024-11-20 13:49:46.626960] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:29:39.603 [2024-11-20 13:49:46.627013] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:29:39.603 [2024-11-20 13:49:46.627173] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:29:39.603 [2024-11-20 13:49:46.627188] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:29:39.603 [2024-11-20 13:49:46.627420] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:29:39.603 [2024-11-20 13:49:46.627556] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:29:39.603 [2024-11-20 13:49:46.627571] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:29:39.603 [2024-11-20 13:49:46.627687] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:39.603 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:39.603 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:29:39.603 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:39.603 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:39.603 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:39.603 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:39.603 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:39.603 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:39.603 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:39.603 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:39.603 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:39.603 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:39.603 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:39.603 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:39.603 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:39.603 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:39.862 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:39.862 "name": "raid_bdev1", 00:29:39.862 "uuid": "dade6eb4-b793-4466-a083-0dae91ecf78c", 00:29:39.862 "strip_size_kb": 64, 00:29:39.862 "state": "online", 00:29:39.862 "raid_level": "concat", 00:29:39.862 "superblock": true, 00:29:39.862 "num_base_bdevs": 4, 00:29:39.862 "num_base_bdevs_discovered": 4, 00:29:39.862 "num_base_bdevs_operational": 4, 00:29:39.862 "base_bdevs_list": [ 00:29:39.862 { 00:29:39.862 "name": "pt1", 00:29:39.862 "uuid": "00000000-0000-0000-0000-000000000001", 00:29:39.862 "is_configured": true, 00:29:39.862 "data_offset": 2048, 00:29:39.862 "data_size": 63488 00:29:39.862 }, 00:29:39.862 { 00:29:39.862 "name": "pt2", 00:29:39.862 "uuid": "00000000-0000-0000-0000-000000000002", 00:29:39.862 "is_configured": true, 00:29:39.862 "data_offset": 2048, 00:29:39.862 "data_size": 63488 00:29:39.862 }, 00:29:39.862 { 00:29:39.862 "name": "pt3", 00:29:39.862 "uuid": "00000000-0000-0000-0000-000000000003", 00:29:39.862 "is_configured": true, 00:29:39.862 "data_offset": 2048, 00:29:39.862 "data_size": 63488 00:29:39.862 }, 00:29:39.862 { 00:29:39.862 "name": "pt4", 00:29:39.862 "uuid": "00000000-0000-0000-0000-000000000004", 00:29:39.862 "is_configured": true, 00:29:39.862 "data_offset": 2048, 00:29:39.862 "data_size": 63488 00:29:39.862 } 00:29:39.862 ] 00:29:39.862 }' 00:29:39.862 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:39.862 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:40.121 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:29:40.121 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:29:40.121 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:29:40.121 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:29:40.121 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:29:40.121 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:29:40.121 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:29:40.121 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:29:40.121 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:40.121 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:40.121 [2024-11-20 13:49:46.961669] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:40.121 13:49:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:40.121 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:40.121 "name": "raid_bdev1", 00:29:40.121 "aliases": [ 00:29:40.121 "dade6eb4-b793-4466-a083-0dae91ecf78c" 00:29:40.121 ], 00:29:40.121 "product_name": "Raid Volume", 00:29:40.121 "block_size": 512, 00:29:40.121 "num_blocks": 253952, 00:29:40.121 "uuid": "dade6eb4-b793-4466-a083-0dae91ecf78c", 00:29:40.121 "assigned_rate_limits": { 00:29:40.121 "rw_ios_per_sec": 0, 00:29:40.121 "rw_mbytes_per_sec": 0, 00:29:40.121 "r_mbytes_per_sec": 0, 00:29:40.121 "w_mbytes_per_sec": 0 00:29:40.121 }, 00:29:40.121 "claimed": false, 00:29:40.121 "zoned": false, 00:29:40.121 "supported_io_types": { 00:29:40.121 "read": true, 00:29:40.121 "write": true, 00:29:40.121 "unmap": true, 00:29:40.121 "flush": true, 00:29:40.121 "reset": true, 00:29:40.121 "nvme_admin": false, 00:29:40.121 "nvme_io": false, 00:29:40.121 "nvme_io_md": false, 00:29:40.121 "write_zeroes": true, 00:29:40.121 "zcopy": false, 00:29:40.121 "get_zone_info": false, 00:29:40.121 "zone_management": false, 00:29:40.121 "zone_append": false, 00:29:40.121 "compare": false, 00:29:40.121 "compare_and_write": false, 00:29:40.121 "abort": false, 00:29:40.121 "seek_hole": false, 00:29:40.121 "seek_data": false, 00:29:40.121 "copy": false, 00:29:40.121 "nvme_iov_md": false 00:29:40.121 }, 00:29:40.121 "memory_domains": [ 00:29:40.121 { 00:29:40.121 "dma_device_id": "system", 00:29:40.121 "dma_device_type": 1 00:29:40.121 }, 00:29:40.121 { 00:29:40.121 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:40.121 "dma_device_type": 2 00:29:40.121 }, 00:29:40.121 { 00:29:40.121 "dma_device_id": "system", 00:29:40.121 "dma_device_type": 1 00:29:40.121 }, 00:29:40.121 { 00:29:40.121 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:40.121 "dma_device_type": 2 00:29:40.121 }, 00:29:40.121 { 00:29:40.121 "dma_device_id": "system", 00:29:40.121 "dma_device_type": 1 00:29:40.121 }, 00:29:40.121 { 00:29:40.121 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:40.121 "dma_device_type": 2 00:29:40.121 }, 00:29:40.121 { 00:29:40.121 "dma_device_id": "system", 00:29:40.121 "dma_device_type": 1 00:29:40.121 }, 00:29:40.121 { 00:29:40.121 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:40.121 "dma_device_type": 2 00:29:40.121 } 00:29:40.121 ], 00:29:40.121 "driver_specific": { 00:29:40.121 "raid": { 00:29:40.121 "uuid": "dade6eb4-b793-4466-a083-0dae91ecf78c", 00:29:40.121 "strip_size_kb": 64, 00:29:40.121 "state": "online", 00:29:40.121 "raid_level": "concat", 00:29:40.121 "superblock": true, 00:29:40.121 "num_base_bdevs": 4, 00:29:40.121 "num_base_bdevs_discovered": 4, 00:29:40.121 "num_base_bdevs_operational": 4, 00:29:40.121 "base_bdevs_list": [ 00:29:40.121 { 00:29:40.121 "name": "pt1", 00:29:40.121 "uuid": "00000000-0000-0000-0000-000000000001", 00:29:40.121 "is_configured": true, 00:29:40.121 "data_offset": 2048, 00:29:40.121 "data_size": 63488 00:29:40.121 }, 00:29:40.121 { 00:29:40.121 "name": "pt2", 00:29:40.121 "uuid": "00000000-0000-0000-0000-000000000002", 00:29:40.121 "is_configured": true, 00:29:40.121 "data_offset": 2048, 00:29:40.121 "data_size": 63488 00:29:40.121 }, 00:29:40.121 { 00:29:40.121 "name": "pt3", 00:29:40.121 "uuid": "00000000-0000-0000-0000-000000000003", 00:29:40.121 "is_configured": true, 00:29:40.121 "data_offset": 2048, 00:29:40.121 "data_size": 63488 00:29:40.121 }, 00:29:40.121 { 00:29:40.121 "name": "pt4", 00:29:40.121 "uuid": "00000000-0000-0000-0000-000000000004", 00:29:40.121 "is_configured": true, 00:29:40.121 "data_offset": 2048, 00:29:40.121 "data_size": 63488 00:29:40.121 } 00:29:40.121 ] 00:29:40.121 } 00:29:40.121 } 00:29:40.121 }' 00:29:40.121 13:49:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:29:40.121 pt2 00:29:40.121 pt3 00:29:40.121 pt4' 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:40.121 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:40.122 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:40.122 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:40.122 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:40.122 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:40.122 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:29:40.122 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:40.122 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:40.122 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:29:40.122 [2024-11-20 13:49:47.177717] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=dade6eb4-b793-4466-a083-0dae91ecf78c 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z dade6eb4-b793-4466-a083-0dae91ecf78c ']' 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:40.380 [2024-11-20 13:49:47.205354] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:29:40.380 [2024-11-20 13:49:47.205385] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:29:40.380 [2024-11-20 13:49:47.205465] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:40.380 [2024-11-20 13:49:47.205538] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:29:40.380 [2024-11-20 13:49:47.205560] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:29:40.380 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:40.381 [2024-11-20 13:49:47.313390] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:29:40.381 [2024-11-20 13:49:47.315133] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:29:40.381 [2024-11-20 13:49:47.315184] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:29:40.381 [2024-11-20 13:49:47.315215] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:29:40.381 [2024-11-20 13:49:47.315261] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:29:40.381 [2024-11-20 13:49:47.315315] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:29:40.381 [2024-11-20 13:49:47.315332] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:29:40.381 [2024-11-20 13:49:47.315349] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:29:40.381 [2024-11-20 13:49:47.315360] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:29:40.381 [2024-11-20 13:49:47.315371] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:29:40.381 request: 00:29:40.381 { 00:29:40.381 "name": "raid_bdev1", 00:29:40.381 "raid_level": "concat", 00:29:40.381 "base_bdevs": [ 00:29:40.381 "malloc1", 00:29:40.381 "malloc2", 00:29:40.381 "malloc3", 00:29:40.381 "malloc4" 00:29:40.381 ], 00:29:40.381 "strip_size_kb": 64, 00:29:40.381 "superblock": false, 00:29:40.381 "method": "bdev_raid_create", 00:29:40.381 "req_id": 1 00:29:40.381 } 00:29:40.381 Got JSON-RPC error response 00:29:40.381 response: 00:29:40.381 { 00:29:40.381 "code": -17, 00:29:40.381 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:29:40.381 } 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:40.381 [2024-11-20 13:49:47.353351] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:29:40.381 [2024-11-20 13:49:47.353402] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:40.381 [2024-11-20 13:49:47.353421] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:29:40.381 [2024-11-20 13:49:47.353430] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:40.381 [2024-11-20 13:49:47.355424] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:40.381 [2024-11-20 13:49:47.355459] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:29:40.381 [2024-11-20 13:49:47.355529] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:29:40.381 [2024-11-20 13:49:47.355578] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:29:40.381 pt1 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 4 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:40.381 "name": "raid_bdev1", 00:29:40.381 "uuid": "dade6eb4-b793-4466-a083-0dae91ecf78c", 00:29:40.381 "strip_size_kb": 64, 00:29:40.381 "state": "configuring", 00:29:40.381 "raid_level": "concat", 00:29:40.381 "superblock": true, 00:29:40.381 "num_base_bdevs": 4, 00:29:40.381 "num_base_bdevs_discovered": 1, 00:29:40.381 "num_base_bdevs_operational": 4, 00:29:40.381 "base_bdevs_list": [ 00:29:40.381 { 00:29:40.381 "name": "pt1", 00:29:40.381 "uuid": "00000000-0000-0000-0000-000000000001", 00:29:40.381 "is_configured": true, 00:29:40.381 "data_offset": 2048, 00:29:40.381 "data_size": 63488 00:29:40.381 }, 00:29:40.381 { 00:29:40.381 "name": null, 00:29:40.381 "uuid": "00000000-0000-0000-0000-000000000002", 00:29:40.381 "is_configured": false, 00:29:40.381 "data_offset": 2048, 00:29:40.381 "data_size": 63488 00:29:40.381 }, 00:29:40.381 { 00:29:40.381 "name": null, 00:29:40.381 "uuid": "00000000-0000-0000-0000-000000000003", 00:29:40.381 "is_configured": false, 00:29:40.381 "data_offset": 2048, 00:29:40.381 "data_size": 63488 00:29:40.381 }, 00:29:40.381 { 00:29:40.381 "name": null, 00:29:40.381 "uuid": "00000000-0000-0000-0000-000000000004", 00:29:40.381 "is_configured": false, 00:29:40.381 "data_offset": 2048, 00:29:40.381 "data_size": 63488 00:29:40.381 } 00:29:40.381 ] 00:29:40.381 }' 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:40.381 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:40.640 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:29:40.640 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:29:40.640 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:40.640 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:40.640 [2024-11-20 13:49:47.685449] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:29:40.640 [2024-11-20 13:49:47.685525] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:40.640 [2024-11-20 13:49:47.685546] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:29:40.640 [2024-11-20 13:49:47.685556] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:40.640 [2024-11-20 13:49:47.685968] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:40.640 [2024-11-20 13:49:47.686000] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:29:40.640 [2024-11-20 13:49:47.686076] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:29:40.640 [2024-11-20 13:49:47.686098] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:29:40.640 pt2 00:29:40.640 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:40.640 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:29:40.640 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:40.640 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:40.640 [2024-11-20 13:49:47.693468] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:29:40.899 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:40.899 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 4 00:29:40.899 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:40.899 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:40.899 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:40.899 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:40.899 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:40.899 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:40.899 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:40.899 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:40.899 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:40.899 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:40.899 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:40.899 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:40.899 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:40.899 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:40.899 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:40.899 "name": "raid_bdev1", 00:29:40.899 "uuid": "dade6eb4-b793-4466-a083-0dae91ecf78c", 00:29:40.899 "strip_size_kb": 64, 00:29:40.899 "state": "configuring", 00:29:40.899 "raid_level": "concat", 00:29:40.899 "superblock": true, 00:29:40.899 "num_base_bdevs": 4, 00:29:40.899 "num_base_bdevs_discovered": 1, 00:29:40.899 "num_base_bdevs_operational": 4, 00:29:40.899 "base_bdevs_list": [ 00:29:40.899 { 00:29:40.899 "name": "pt1", 00:29:40.899 "uuid": "00000000-0000-0000-0000-000000000001", 00:29:40.899 "is_configured": true, 00:29:40.899 "data_offset": 2048, 00:29:40.899 "data_size": 63488 00:29:40.899 }, 00:29:40.899 { 00:29:40.899 "name": null, 00:29:40.899 "uuid": "00000000-0000-0000-0000-000000000002", 00:29:40.899 "is_configured": false, 00:29:40.899 "data_offset": 0, 00:29:40.899 "data_size": 63488 00:29:40.899 }, 00:29:40.899 { 00:29:40.899 "name": null, 00:29:40.899 "uuid": "00000000-0000-0000-0000-000000000003", 00:29:40.899 "is_configured": false, 00:29:40.899 "data_offset": 2048, 00:29:40.899 "data_size": 63488 00:29:40.899 }, 00:29:40.899 { 00:29:40.899 "name": null, 00:29:40.899 "uuid": "00000000-0000-0000-0000-000000000004", 00:29:40.899 "is_configured": false, 00:29:40.899 "data_offset": 2048, 00:29:40.899 "data_size": 63488 00:29:40.899 } 00:29:40.899 ] 00:29:40.899 }' 00:29:40.899 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:40.899 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:41.157 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:29:41.157 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:29:41.157 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:29:41.157 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:41.157 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:41.157 [2024-11-20 13:49:47.993521] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:29:41.157 [2024-11-20 13:49:47.993589] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:41.157 [2024-11-20 13:49:47.993608] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:29:41.157 [2024-11-20 13:49:47.993617] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:41.157 [2024-11-20 13:49:47.994043] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:41.157 [2024-11-20 13:49:47.994061] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:29:41.157 [2024-11-20 13:49:47.994136] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:29:41.157 [2024-11-20 13:49:47.994161] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:29:41.157 pt2 00:29:41.157 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:41.157 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:29:41.157 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:29:41.157 13:49:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:29:41.157 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:41.157 13:49:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:41.157 [2024-11-20 13:49:48.001474] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:29:41.157 [2024-11-20 13:49:48.001519] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:41.157 [2024-11-20 13:49:48.001535] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:29:41.157 [2024-11-20 13:49:48.001543] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:41.157 [2024-11-20 13:49:48.001879] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:41.157 [2024-11-20 13:49:48.001900] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:29:41.157 [2024-11-20 13:49:48.001956] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:29:41.157 [2024-11-20 13:49:48.001992] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:29:41.157 pt3 00:29:41.157 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:41.157 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:29:41.157 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:29:41.157 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:29:41.157 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:41.157 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:41.157 [2024-11-20 13:49:48.009449] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:29:41.157 [2024-11-20 13:49:48.009484] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:41.157 [2024-11-20 13:49:48.009498] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:29:41.157 [2024-11-20 13:49:48.009505] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:41.157 [2024-11-20 13:49:48.009810] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:41.157 [2024-11-20 13:49:48.009827] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:29:41.157 [2024-11-20 13:49:48.009876] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:29:41.157 [2024-11-20 13:49:48.009892] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:29:41.157 [2024-11-20 13:49:48.010016] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:29:41.157 [2024-11-20 13:49:48.010030] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:29:41.157 [2024-11-20 13:49:48.010232] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:29:41.157 [2024-11-20 13:49:48.010348] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:29:41.157 [2024-11-20 13:49:48.010361] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:29:41.157 [2024-11-20 13:49:48.010462] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:41.157 pt4 00:29:41.157 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:41.158 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:29:41.158 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:29:41.158 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:29:41.158 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:41.158 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:41.158 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:41.158 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:41.158 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:41.158 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:41.158 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:41.158 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:41.158 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:41.158 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:41.158 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:41.158 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:41.158 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:41.158 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:41.158 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:41.158 "name": "raid_bdev1", 00:29:41.158 "uuid": "dade6eb4-b793-4466-a083-0dae91ecf78c", 00:29:41.158 "strip_size_kb": 64, 00:29:41.158 "state": "online", 00:29:41.158 "raid_level": "concat", 00:29:41.158 "superblock": true, 00:29:41.158 "num_base_bdevs": 4, 00:29:41.158 "num_base_bdevs_discovered": 4, 00:29:41.158 "num_base_bdevs_operational": 4, 00:29:41.158 "base_bdevs_list": [ 00:29:41.158 { 00:29:41.158 "name": "pt1", 00:29:41.158 "uuid": "00000000-0000-0000-0000-000000000001", 00:29:41.158 "is_configured": true, 00:29:41.158 "data_offset": 2048, 00:29:41.158 "data_size": 63488 00:29:41.158 }, 00:29:41.158 { 00:29:41.158 "name": "pt2", 00:29:41.158 "uuid": "00000000-0000-0000-0000-000000000002", 00:29:41.158 "is_configured": true, 00:29:41.158 "data_offset": 2048, 00:29:41.158 "data_size": 63488 00:29:41.158 }, 00:29:41.158 { 00:29:41.158 "name": "pt3", 00:29:41.158 "uuid": "00000000-0000-0000-0000-000000000003", 00:29:41.158 "is_configured": true, 00:29:41.158 "data_offset": 2048, 00:29:41.158 "data_size": 63488 00:29:41.158 }, 00:29:41.158 { 00:29:41.158 "name": "pt4", 00:29:41.158 "uuid": "00000000-0000-0000-0000-000000000004", 00:29:41.158 "is_configured": true, 00:29:41.158 "data_offset": 2048, 00:29:41.158 "data_size": 63488 00:29:41.158 } 00:29:41.158 ] 00:29:41.158 }' 00:29:41.158 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:41.158 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:41.418 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:29:41.418 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:29:41.418 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:29:41.418 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:29:41.418 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:29:41.418 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:29:41.418 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:29:41.418 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:29:41.418 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:41.418 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:41.418 [2024-11-20 13:49:48.337885] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:41.418 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:41.418 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:41.418 "name": "raid_bdev1", 00:29:41.418 "aliases": [ 00:29:41.418 "dade6eb4-b793-4466-a083-0dae91ecf78c" 00:29:41.418 ], 00:29:41.418 "product_name": "Raid Volume", 00:29:41.418 "block_size": 512, 00:29:41.418 "num_blocks": 253952, 00:29:41.418 "uuid": "dade6eb4-b793-4466-a083-0dae91ecf78c", 00:29:41.418 "assigned_rate_limits": { 00:29:41.418 "rw_ios_per_sec": 0, 00:29:41.418 "rw_mbytes_per_sec": 0, 00:29:41.418 "r_mbytes_per_sec": 0, 00:29:41.418 "w_mbytes_per_sec": 0 00:29:41.418 }, 00:29:41.418 "claimed": false, 00:29:41.418 "zoned": false, 00:29:41.418 "supported_io_types": { 00:29:41.418 "read": true, 00:29:41.418 "write": true, 00:29:41.418 "unmap": true, 00:29:41.418 "flush": true, 00:29:41.418 "reset": true, 00:29:41.418 "nvme_admin": false, 00:29:41.418 "nvme_io": false, 00:29:41.418 "nvme_io_md": false, 00:29:41.418 "write_zeroes": true, 00:29:41.418 "zcopy": false, 00:29:41.418 "get_zone_info": false, 00:29:41.418 "zone_management": false, 00:29:41.418 "zone_append": false, 00:29:41.418 "compare": false, 00:29:41.418 "compare_and_write": false, 00:29:41.418 "abort": false, 00:29:41.418 "seek_hole": false, 00:29:41.418 "seek_data": false, 00:29:41.418 "copy": false, 00:29:41.418 "nvme_iov_md": false 00:29:41.418 }, 00:29:41.418 "memory_domains": [ 00:29:41.418 { 00:29:41.418 "dma_device_id": "system", 00:29:41.418 "dma_device_type": 1 00:29:41.418 }, 00:29:41.418 { 00:29:41.418 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:41.418 "dma_device_type": 2 00:29:41.418 }, 00:29:41.418 { 00:29:41.418 "dma_device_id": "system", 00:29:41.418 "dma_device_type": 1 00:29:41.418 }, 00:29:41.418 { 00:29:41.418 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:41.418 "dma_device_type": 2 00:29:41.418 }, 00:29:41.418 { 00:29:41.418 "dma_device_id": "system", 00:29:41.418 "dma_device_type": 1 00:29:41.418 }, 00:29:41.418 { 00:29:41.418 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:41.418 "dma_device_type": 2 00:29:41.418 }, 00:29:41.418 { 00:29:41.418 "dma_device_id": "system", 00:29:41.418 "dma_device_type": 1 00:29:41.418 }, 00:29:41.418 { 00:29:41.418 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:41.418 "dma_device_type": 2 00:29:41.418 } 00:29:41.418 ], 00:29:41.418 "driver_specific": { 00:29:41.418 "raid": { 00:29:41.418 "uuid": "dade6eb4-b793-4466-a083-0dae91ecf78c", 00:29:41.418 "strip_size_kb": 64, 00:29:41.418 "state": "online", 00:29:41.418 "raid_level": "concat", 00:29:41.418 "superblock": true, 00:29:41.418 "num_base_bdevs": 4, 00:29:41.418 "num_base_bdevs_discovered": 4, 00:29:41.418 "num_base_bdevs_operational": 4, 00:29:41.418 "base_bdevs_list": [ 00:29:41.418 { 00:29:41.418 "name": "pt1", 00:29:41.418 "uuid": "00000000-0000-0000-0000-000000000001", 00:29:41.418 "is_configured": true, 00:29:41.418 "data_offset": 2048, 00:29:41.418 "data_size": 63488 00:29:41.418 }, 00:29:41.418 { 00:29:41.418 "name": "pt2", 00:29:41.418 "uuid": "00000000-0000-0000-0000-000000000002", 00:29:41.418 "is_configured": true, 00:29:41.418 "data_offset": 2048, 00:29:41.418 "data_size": 63488 00:29:41.418 }, 00:29:41.418 { 00:29:41.418 "name": "pt3", 00:29:41.418 "uuid": "00000000-0000-0000-0000-000000000003", 00:29:41.418 "is_configured": true, 00:29:41.418 "data_offset": 2048, 00:29:41.418 "data_size": 63488 00:29:41.418 }, 00:29:41.418 { 00:29:41.418 "name": "pt4", 00:29:41.418 "uuid": "00000000-0000-0000-0000-000000000004", 00:29:41.418 "is_configured": true, 00:29:41.418 "data_offset": 2048, 00:29:41.418 "data_size": 63488 00:29:41.418 } 00:29:41.418 ] 00:29:41.418 } 00:29:41.418 } 00:29:41.418 }' 00:29:41.418 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:29:41.418 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:29:41.418 pt2 00:29:41.418 pt3 00:29:41.418 pt4' 00:29:41.418 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:41.418 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:29:41.419 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:41.419 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:29:41.419 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:41.419 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:41.419 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:41.419 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:41.419 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:41.419 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:41.419 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:41.419 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:29:41.419 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:41.419 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:41.419 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:41.419 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:41.679 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:41.679 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:41.679 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:41.679 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:29:41.679 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:41.679 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:41.679 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:41.679 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:41.679 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:41.679 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:41.679 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:41.679 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:41.679 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:29:41.679 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:41.679 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:41.679 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:41.679 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:41.679 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:41.679 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:29:41.679 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:29:41.679 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:41.680 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:41.680 [2024-11-20 13:49:48.541888] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:41.680 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:41.680 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' dade6eb4-b793-4466-a083-0dae91ecf78c '!=' dade6eb4-b793-4466-a083-0dae91ecf78c ']' 00:29:41.680 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:29:41.680 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:29:41.680 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:29:41.680 13:49:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 70681 00:29:41.680 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 70681 ']' 00:29:41.680 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 70681 00:29:41.680 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:29:41.680 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:29:41.680 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70681 00:29:41.680 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:29:41.680 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:29:41.680 killing process with pid 70681 00:29:41.680 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70681' 00:29:41.680 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 70681 00:29:41.680 13:49:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 70681 00:29:41.680 [2024-11-20 13:49:48.581327] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:29:41.680 [2024-11-20 13:49:48.581417] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:41.680 [2024-11-20 13:49:48.581508] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:29:41.680 [2024-11-20 13:49:48.581517] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:29:41.937 [2024-11-20 13:49:48.793311] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:29:42.503 13:49:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:29:42.503 00:29:42.503 real 0m3.833s 00:29:42.503 user 0m5.475s 00:29:42.503 sys 0m0.669s 00:29:42.503 13:49:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:29:42.503 13:49:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:42.503 ************************************ 00:29:42.503 END TEST raid_superblock_test 00:29:42.503 ************************************ 00:29:42.503 13:49:49 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 4 read 00:29:42.503 13:49:49 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:29:42.503 13:49:49 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:29:42.503 13:49:49 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:29:42.503 ************************************ 00:29:42.503 START TEST raid_read_error_test 00:29:42.503 ************************************ 00:29:42.503 13:49:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 4 read 00:29:42.503 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:29:42.503 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:29:42.503 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:29:42.503 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:29:42.503 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:29:42.503 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:29:42.503 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:29:42.503 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:29:42.503 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:29:42.503 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:29:42.503 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:29:42.503 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:29:42.503 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:29:42.503 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:29:42.503 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:29:42.503 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:29:42.503 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:29:42.503 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:29:42.503 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:29:42.503 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:29:42.503 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:29:42.503 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:29:42.503 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:29:42.504 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:29:42.504 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:29:42.504 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:29:42.504 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:29:42.504 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:29:42.504 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.Z886EodWFN 00:29:42.504 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=70929 00:29:42.504 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 70929 00:29:42.504 13:49:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 70929 ']' 00:29:42.504 13:49:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:42.504 13:49:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:29:42.504 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:42.504 13:49:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:42.504 13:49:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:29:42.504 13:49:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:42.504 13:49:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:29:42.504 [2024-11-20 13:49:49.538291] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:29:42.504 [2024-11-20 13:49:49.538425] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70929 ] 00:29:42.777 [2024-11-20 13:49:49.697377] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:42.777 [2024-11-20 13:49:49.798199] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:29:43.036 [2024-11-20 13:49:49.919904] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:43.036 [2024-11-20 13:49:49.919962] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:43.295 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:29:43.295 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:29:43.295 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:29:43.295 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:29:43.295 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:43.295 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:43.554 BaseBdev1_malloc 00:29:43.554 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:43.554 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:29:43.554 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:43.554 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:43.554 true 00:29:43.554 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:43.554 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:29:43.554 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:43.554 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:43.554 [2024-11-20 13:49:50.367714] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:29:43.554 [2024-11-20 13:49:50.367774] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:43.554 [2024-11-20 13:49:50.367794] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:29:43.554 [2024-11-20 13:49:50.367804] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:43.554 [2024-11-20 13:49:50.369786] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:43.554 [2024-11-20 13:49:50.369821] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:29:43.554 BaseBdev1 00:29:43.554 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:43.554 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:29:43.554 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:29:43.554 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:43.554 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:43.554 BaseBdev2_malloc 00:29:43.554 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:43.554 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:29:43.554 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:43.554 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:43.554 true 00:29:43.554 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:43.554 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:43.555 [2024-11-20 13:49:50.409608] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:29:43.555 [2024-11-20 13:49:50.409660] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:43.555 [2024-11-20 13:49:50.409674] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:29:43.555 [2024-11-20 13:49:50.409684] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:43.555 [2024-11-20 13:49:50.411579] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:43.555 [2024-11-20 13:49:50.411611] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:29:43.555 BaseBdev2 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:43.555 BaseBdev3_malloc 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:43.555 true 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:43.555 [2024-11-20 13:49:50.466410] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:29:43.555 [2024-11-20 13:49:50.466462] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:43.555 [2024-11-20 13:49:50.466479] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:29:43.555 [2024-11-20 13:49:50.466488] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:43.555 [2024-11-20 13:49:50.468393] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:43.555 [2024-11-20 13:49:50.468425] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:29:43.555 BaseBdev3 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:43.555 BaseBdev4_malloc 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:43.555 true 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:43.555 [2024-11-20 13:49:50.507968] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:29:43.555 [2024-11-20 13:49:50.508028] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:43.555 [2024-11-20 13:49:50.508044] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:29:43.555 [2024-11-20 13:49:50.508053] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:43.555 [2024-11-20 13:49:50.509923] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:43.555 [2024-11-20 13:49:50.509958] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:29:43.555 BaseBdev4 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:43.555 [2024-11-20 13:49:50.516050] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:29:43.555 [2024-11-20 13:49:50.517692] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:29:43.555 [2024-11-20 13:49:50.517761] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:29:43.555 [2024-11-20 13:49:50.517817] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:29:43.555 [2024-11-20 13:49:50.518016] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008580 00:29:43.555 [2024-11-20 13:49:50.518032] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:29:43.555 [2024-11-20 13:49:50.518243] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000068a0 00:29:43.555 [2024-11-20 13:49:50.518376] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008580 00:29:43.555 [2024-11-20 13:49:50.518390] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008580 00:29:43.555 [2024-11-20 13:49:50.518508] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:43.555 "name": "raid_bdev1", 00:29:43.555 "uuid": "929e5772-d75c-41a0-92ba-941f5a2f18c5", 00:29:43.555 "strip_size_kb": 64, 00:29:43.555 "state": "online", 00:29:43.555 "raid_level": "concat", 00:29:43.555 "superblock": true, 00:29:43.555 "num_base_bdevs": 4, 00:29:43.555 "num_base_bdevs_discovered": 4, 00:29:43.555 "num_base_bdevs_operational": 4, 00:29:43.555 "base_bdevs_list": [ 00:29:43.555 { 00:29:43.555 "name": "BaseBdev1", 00:29:43.555 "uuid": "0bc695ee-16d3-5189-9199-ed5f4d754b4b", 00:29:43.555 "is_configured": true, 00:29:43.555 "data_offset": 2048, 00:29:43.555 "data_size": 63488 00:29:43.555 }, 00:29:43.555 { 00:29:43.555 "name": "BaseBdev2", 00:29:43.555 "uuid": "f254371b-229f-5788-9fb8-cf2bcec5dd94", 00:29:43.555 "is_configured": true, 00:29:43.555 "data_offset": 2048, 00:29:43.555 "data_size": 63488 00:29:43.555 }, 00:29:43.555 { 00:29:43.555 "name": "BaseBdev3", 00:29:43.555 "uuid": "f472c718-fc5e-52f8-8e56-998e7885f607", 00:29:43.555 "is_configured": true, 00:29:43.555 "data_offset": 2048, 00:29:43.555 "data_size": 63488 00:29:43.555 }, 00:29:43.555 { 00:29:43.555 "name": "BaseBdev4", 00:29:43.555 "uuid": "41812df6-6335-5359-bb62-a75aacf7f0db", 00:29:43.555 "is_configured": true, 00:29:43.555 "data_offset": 2048, 00:29:43.555 "data_size": 63488 00:29:43.555 } 00:29:43.555 ] 00:29:43.555 }' 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:43.555 13:49:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:43.817 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:29:43.817 13:49:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:29:44.075 [2024-11-20 13:49:50.933032] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006a40 00:29:45.011 13:49:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:29:45.011 13:49:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:45.011 13:49:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:45.011 13:49:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:45.011 13:49:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:29:45.011 13:49:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:29:45.011 13:49:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:29:45.011 13:49:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:29:45.011 13:49:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:45.011 13:49:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:45.011 13:49:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:45.011 13:49:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:45.011 13:49:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:45.011 13:49:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:45.011 13:49:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:45.011 13:49:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:45.011 13:49:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:45.011 13:49:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:45.011 13:49:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:45.011 13:49:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:45.011 13:49:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:45.011 13:49:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:45.011 13:49:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:45.011 "name": "raid_bdev1", 00:29:45.011 "uuid": "929e5772-d75c-41a0-92ba-941f5a2f18c5", 00:29:45.011 "strip_size_kb": 64, 00:29:45.011 "state": "online", 00:29:45.011 "raid_level": "concat", 00:29:45.011 "superblock": true, 00:29:45.011 "num_base_bdevs": 4, 00:29:45.011 "num_base_bdevs_discovered": 4, 00:29:45.011 "num_base_bdevs_operational": 4, 00:29:45.011 "base_bdevs_list": [ 00:29:45.011 { 00:29:45.011 "name": "BaseBdev1", 00:29:45.011 "uuid": "0bc695ee-16d3-5189-9199-ed5f4d754b4b", 00:29:45.011 "is_configured": true, 00:29:45.011 "data_offset": 2048, 00:29:45.011 "data_size": 63488 00:29:45.011 }, 00:29:45.011 { 00:29:45.011 "name": "BaseBdev2", 00:29:45.011 "uuid": "f254371b-229f-5788-9fb8-cf2bcec5dd94", 00:29:45.011 "is_configured": true, 00:29:45.011 "data_offset": 2048, 00:29:45.011 "data_size": 63488 00:29:45.011 }, 00:29:45.011 { 00:29:45.011 "name": "BaseBdev3", 00:29:45.011 "uuid": "f472c718-fc5e-52f8-8e56-998e7885f607", 00:29:45.011 "is_configured": true, 00:29:45.011 "data_offset": 2048, 00:29:45.011 "data_size": 63488 00:29:45.011 }, 00:29:45.011 { 00:29:45.011 "name": "BaseBdev4", 00:29:45.011 "uuid": "41812df6-6335-5359-bb62-a75aacf7f0db", 00:29:45.011 "is_configured": true, 00:29:45.011 "data_offset": 2048, 00:29:45.011 "data_size": 63488 00:29:45.011 } 00:29:45.011 ] 00:29:45.011 }' 00:29:45.011 13:49:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:45.011 13:49:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:45.269 13:49:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:29:45.269 13:49:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:45.269 13:49:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:45.269 [2024-11-20 13:49:52.170446] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:29:45.269 [2024-11-20 13:49:52.170487] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:29:45.269 [2024-11-20 13:49:52.172996] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:45.269 [2024-11-20 13:49:52.173063] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:45.269 [2024-11-20 13:49:52.173105] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:29:45.269 [2024-11-20 13:49:52.173117] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state offline 00:29:45.269 13:49:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:45.269 13:49:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 70929 00:29:45.269 13:49:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 70929 ']' 00:29:45.269 13:49:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 70929 00:29:45.269 { 00:29:45.269 "results": [ 00:29:45.269 { 00:29:45.269 "job": "raid_bdev1", 00:29:45.269 "core_mask": "0x1", 00:29:45.269 "workload": "randrw", 00:29:45.269 "percentage": 50, 00:29:45.269 "status": "finished", 00:29:45.269 "queue_depth": 1, 00:29:45.269 "io_size": 131072, 00:29:45.269 "runtime": 1.23573, 00:29:45.269 "iops": 16212.279381418271, 00:29:45.269 "mibps": 2026.534922677284, 00:29:45.269 "io_failed": 1, 00:29:45.269 "io_timeout": 0, 00:29:45.269 "avg_latency_us": 84.84164650323473, 00:29:45.269 "min_latency_us": 26.78153846153846, 00:29:45.269 "max_latency_us": 1348.5292307692307 00:29:45.269 } 00:29:45.269 ], 00:29:45.269 "core_count": 1 00:29:45.269 } 00:29:45.269 13:49:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:29:45.270 13:49:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:29:45.270 13:49:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70929 00:29:45.270 killing process with pid 70929 00:29:45.270 13:49:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:29:45.270 13:49:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:29:45.270 13:49:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70929' 00:29:45.270 13:49:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 70929 00:29:45.270 13:49:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 70929 00:29:45.270 [2024-11-20 13:49:52.200193] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:29:45.527 [2024-11-20 13:49:52.370235] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:29:46.093 13:49:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:29:46.093 13:49:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.Z886EodWFN 00:29:46.093 13:49:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:29:46.093 13:49:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.81 00:29:46.093 13:49:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:29:46.093 13:49:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:29:46.093 13:49:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:29:46.093 13:49:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.81 != \0\.\0\0 ]] 00:29:46.093 00:29:46.093 real 0m3.574s 00:29:46.093 user 0m4.172s 00:29:46.093 sys 0m0.435s 00:29:46.094 13:49:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:29:46.094 ************************************ 00:29:46.094 END TEST raid_read_error_test 00:29:46.094 ************************************ 00:29:46.094 13:49:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:46.094 13:49:53 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 4 write 00:29:46.094 13:49:53 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:29:46.094 13:49:53 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:29:46.094 13:49:53 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:29:46.094 ************************************ 00:29:46.094 START TEST raid_write_error_test 00:29:46.094 ************************************ 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 4 write 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.umQNu1sLd8 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=71058 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 71058 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 71058 ']' 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:29:46.094 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:29:46.094 13:49:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:46.416 [2024-11-20 13:49:53.154170] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:29:46.416 [2024-11-20 13:49:53.154298] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71058 ] 00:29:46.416 [2024-11-20 13:49:53.311647] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:46.416 [2024-11-20 13:49:53.413822] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:29:46.684 [2024-11-20 13:49:53.536958] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:46.684 [2024-11-20 13:49:53.537020] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:46.942 13:49:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:29:46.942 13:49:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:29:46.942 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:29:46.942 13:49:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:29:46.942 13:49:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:46.942 13:49:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:47.200 BaseBdev1_malloc 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:47.200 true 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:47.200 [2024-11-20 13:49:54.032101] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:29:47.200 [2024-11-20 13:49:54.032161] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:47.200 [2024-11-20 13:49:54.032178] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:29:47.200 [2024-11-20 13:49:54.032189] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:47.200 [2024-11-20 13:49:54.034074] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:47.200 [2024-11-20 13:49:54.034108] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:29:47.200 BaseBdev1 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:47.200 BaseBdev2_malloc 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:47.200 true 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:47.200 [2024-11-20 13:49:54.073807] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:29:47.200 [2024-11-20 13:49:54.073855] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:47.200 [2024-11-20 13:49:54.073869] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:29:47.200 [2024-11-20 13:49:54.073878] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:47.200 [2024-11-20 13:49:54.075718] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:47.200 [2024-11-20 13:49:54.075751] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:29:47.200 BaseBdev2 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:47.200 BaseBdev3_malloc 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:47.200 true 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:47.200 [2024-11-20 13:49:54.132352] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:29:47.200 [2024-11-20 13:49:54.132403] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:47.200 [2024-11-20 13:49:54.132418] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:29:47.200 [2024-11-20 13:49:54.132434] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:47.200 [2024-11-20 13:49:54.134317] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:47.200 [2024-11-20 13:49:54.134350] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:29:47.200 BaseBdev3 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:47.200 BaseBdev4_malloc 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:47.200 true 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:47.200 [2024-11-20 13:49:54.174191] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:29:47.200 [2024-11-20 13:49:54.174240] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:47.200 [2024-11-20 13:49:54.174255] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:29:47.200 [2024-11-20 13:49:54.174264] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:47.200 [2024-11-20 13:49:54.176120] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:47.200 [2024-11-20 13:49:54.176153] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:29:47.200 BaseBdev4 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:47.200 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:47.200 [2024-11-20 13:49:54.182262] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:29:47.200 [2024-11-20 13:49:54.183877] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:29:47.200 [2024-11-20 13:49:54.183947] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:29:47.200 [2024-11-20 13:49:54.184014] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:29:47.201 [2024-11-20 13:49:54.184201] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008580 00:29:47.201 [2024-11-20 13:49:54.184217] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:29:47.201 [2024-11-20 13:49:54.184425] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000068a0 00:29:47.201 [2024-11-20 13:49:54.184577] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008580 00:29:47.201 [2024-11-20 13:49:54.184614] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008580 00:29:47.201 [2024-11-20 13:49:54.184736] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:47.201 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:47.201 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:29:47.201 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:47.201 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:47.201 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:47.201 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:47.201 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:47.201 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:47.201 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:47.201 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:47.201 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:47.201 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:47.201 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:47.201 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:47.201 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:47.201 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:47.201 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:47.201 "name": "raid_bdev1", 00:29:47.201 "uuid": "1327632f-6c93-4032-b04b-843d55fbe96e", 00:29:47.201 "strip_size_kb": 64, 00:29:47.201 "state": "online", 00:29:47.201 "raid_level": "concat", 00:29:47.201 "superblock": true, 00:29:47.201 "num_base_bdevs": 4, 00:29:47.201 "num_base_bdevs_discovered": 4, 00:29:47.201 "num_base_bdevs_operational": 4, 00:29:47.201 "base_bdevs_list": [ 00:29:47.201 { 00:29:47.201 "name": "BaseBdev1", 00:29:47.201 "uuid": "b414cd96-21ec-59af-9106-30a701f3691e", 00:29:47.201 "is_configured": true, 00:29:47.201 "data_offset": 2048, 00:29:47.201 "data_size": 63488 00:29:47.201 }, 00:29:47.201 { 00:29:47.201 "name": "BaseBdev2", 00:29:47.201 "uuid": "2629dec6-5e19-5079-adb4-a01ae14edd87", 00:29:47.201 "is_configured": true, 00:29:47.201 "data_offset": 2048, 00:29:47.201 "data_size": 63488 00:29:47.201 }, 00:29:47.201 { 00:29:47.201 "name": "BaseBdev3", 00:29:47.201 "uuid": "f0cca313-a656-54e0-a9d7-5b02929489be", 00:29:47.201 "is_configured": true, 00:29:47.201 "data_offset": 2048, 00:29:47.201 "data_size": 63488 00:29:47.201 }, 00:29:47.201 { 00:29:47.201 "name": "BaseBdev4", 00:29:47.201 "uuid": "2a20d5a7-db4e-5907-86a0-8a01740c7e4b", 00:29:47.201 "is_configured": true, 00:29:47.201 "data_offset": 2048, 00:29:47.201 "data_size": 63488 00:29:47.201 } 00:29:47.201 ] 00:29:47.201 }' 00:29:47.201 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:47.201 13:49:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:47.458 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:29:47.458 13:49:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:29:47.715 [2024-11-20 13:49:54.595252] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006a40 00:29:48.648 13:49:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:29:48.648 13:49:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:48.648 13:49:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:48.648 13:49:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:48.648 13:49:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:29:48.648 13:49:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:29:48.648 13:49:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:29:48.648 13:49:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:29:48.648 13:49:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:48.648 13:49:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:48.648 13:49:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:29:48.648 13:49:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:48.648 13:49:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:48.648 13:49:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:48.648 13:49:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:48.648 13:49:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:48.648 13:49:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:48.648 13:49:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:48.648 13:49:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:48.648 13:49:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:48.648 13:49:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:48.648 13:49:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:48.648 13:49:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:48.648 "name": "raid_bdev1", 00:29:48.648 "uuid": "1327632f-6c93-4032-b04b-843d55fbe96e", 00:29:48.648 "strip_size_kb": 64, 00:29:48.648 "state": "online", 00:29:48.648 "raid_level": "concat", 00:29:48.648 "superblock": true, 00:29:48.648 "num_base_bdevs": 4, 00:29:48.648 "num_base_bdevs_discovered": 4, 00:29:48.648 "num_base_bdevs_operational": 4, 00:29:48.648 "base_bdevs_list": [ 00:29:48.648 { 00:29:48.648 "name": "BaseBdev1", 00:29:48.648 "uuid": "b414cd96-21ec-59af-9106-30a701f3691e", 00:29:48.648 "is_configured": true, 00:29:48.648 "data_offset": 2048, 00:29:48.648 "data_size": 63488 00:29:48.648 }, 00:29:48.648 { 00:29:48.648 "name": "BaseBdev2", 00:29:48.648 "uuid": "2629dec6-5e19-5079-adb4-a01ae14edd87", 00:29:48.648 "is_configured": true, 00:29:48.648 "data_offset": 2048, 00:29:48.648 "data_size": 63488 00:29:48.648 }, 00:29:48.648 { 00:29:48.648 "name": "BaseBdev3", 00:29:48.648 "uuid": "f0cca313-a656-54e0-a9d7-5b02929489be", 00:29:48.648 "is_configured": true, 00:29:48.648 "data_offset": 2048, 00:29:48.648 "data_size": 63488 00:29:48.648 }, 00:29:48.648 { 00:29:48.648 "name": "BaseBdev4", 00:29:48.648 "uuid": "2a20d5a7-db4e-5907-86a0-8a01740c7e4b", 00:29:48.648 "is_configured": true, 00:29:48.648 "data_offset": 2048, 00:29:48.648 "data_size": 63488 00:29:48.648 } 00:29:48.648 ] 00:29:48.648 }' 00:29:48.648 13:49:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:48.648 13:49:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:48.906 13:49:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:29:48.906 13:49:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:48.906 13:49:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:48.906 [2024-11-20 13:49:55.828359] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:29:48.906 [2024-11-20 13:49:55.828399] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:29:48.906 [2024-11-20 13:49:55.830910] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:48.906 [2024-11-20 13:49:55.830973] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:48.906 [2024-11-20 13:49:55.831025] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:29:48.906 [2024-11-20 13:49:55.831036] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state offline 00:29:48.906 13:49:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:48.906 { 00:29:48.906 "results": [ 00:29:48.906 { 00:29:48.906 "job": "raid_bdev1", 00:29:48.906 "core_mask": "0x1", 00:29:48.906 "workload": "randrw", 00:29:48.906 "percentage": 50, 00:29:48.906 "status": "finished", 00:29:48.906 "queue_depth": 1, 00:29:48.906 "io_size": 131072, 00:29:48.906 "runtime": 1.231456, 00:29:48.906 "iops": 16413.903541823664, 00:29:48.906 "mibps": 2051.737942727958, 00:29:48.906 "io_failed": 1, 00:29:48.906 "io_timeout": 0, 00:29:48.906 "avg_latency_us": 83.75872807117688, 00:29:48.906 "min_latency_us": 25.993846153846153, 00:29:48.906 "max_latency_us": 1310.72 00:29:48.906 } 00:29:48.906 ], 00:29:48.906 "core_count": 1 00:29:48.906 } 00:29:48.906 13:49:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 71058 00:29:48.906 13:49:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 71058 ']' 00:29:48.906 13:49:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 71058 00:29:48.906 13:49:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:29:48.906 13:49:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:29:48.906 13:49:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71058 00:29:48.906 killing process with pid 71058 00:29:48.906 13:49:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:29:48.906 13:49:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:29:48.906 13:49:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71058' 00:29:48.906 13:49:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 71058 00:29:48.906 13:49:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 71058 00:29:48.906 [2024-11-20 13:49:55.855759] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:29:49.164 [2024-11-20 13:49:56.024413] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:29:49.729 13:49:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.umQNu1sLd8 00:29:49.729 13:49:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:29:49.729 13:49:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:29:49.729 13:49:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.81 00:29:49.729 13:49:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:29:49.729 13:49:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:29:49.730 13:49:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:29:49.730 13:49:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.81 != \0\.\0\0 ]] 00:29:49.730 00:29:49.730 real 0m3.601s 00:29:49.730 user 0m4.261s 00:29:49.730 sys 0m0.420s 00:29:49.730 13:49:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:29:49.730 13:49:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:29:49.730 ************************************ 00:29:49.730 END TEST raid_write_error_test 00:29:49.730 ************************************ 00:29:49.730 13:49:56 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:29:49.730 13:49:56 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 4 false 00:29:49.730 13:49:56 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:29:49.730 13:49:56 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:29:49.730 13:49:56 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:29:49.730 ************************************ 00:29:49.730 START TEST raid_state_function_test 00:29:49.730 ************************************ 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 4 false 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=71196 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 71196' 00:29:49.730 Process raid pid: 71196 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 71196 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 71196 ']' 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:29:49.730 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:49.730 13:49:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:29:49.987 [2024-11-20 13:49:56.795940] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:29:49.988 [2024-11-20 13:49:56.796080] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:29:49.988 [2024-11-20 13:49:56.954443] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:50.245 [2024-11-20 13:49:57.055975] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:29:50.245 [2024-11-20 13:49:57.177018] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:50.245 [2024-11-20 13:49:57.177056] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:50.810 13:49:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:29:50.811 13:49:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:29:50.811 13:49:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:29:50.811 13:49:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:50.811 13:49:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:50.811 [2024-11-20 13:49:57.675085] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:29:50.811 [2024-11-20 13:49:57.675146] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:29:50.811 [2024-11-20 13:49:57.675156] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:29:50.811 [2024-11-20 13:49:57.675164] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:29:50.811 [2024-11-20 13:49:57.675170] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:29:50.811 [2024-11-20 13:49:57.675177] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:29:50.811 [2024-11-20 13:49:57.675182] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:29:50.811 [2024-11-20 13:49:57.675189] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:29:50.811 13:49:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:50.811 13:49:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:29:50.811 13:49:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:50.811 13:49:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:50.811 13:49:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:29:50.811 13:49:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:29:50.811 13:49:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:50.811 13:49:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:50.811 13:49:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:50.811 13:49:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:50.811 13:49:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:50.811 13:49:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:50.811 13:49:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:50.811 13:49:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:50.811 13:49:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:50.811 13:49:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:50.811 13:49:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:50.811 "name": "Existed_Raid", 00:29:50.811 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:50.811 "strip_size_kb": 0, 00:29:50.811 "state": "configuring", 00:29:50.811 "raid_level": "raid1", 00:29:50.811 "superblock": false, 00:29:50.811 "num_base_bdevs": 4, 00:29:50.811 "num_base_bdevs_discovered": 0, 00:29:50.811 "num_base_bdevs_operational": 4, 00:29:50.811 "base_bdevs_list": [ 00:29:50.811 { 00:29:50.811 "name": "BaseBdev1", 00:29:50.811 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:50.811 "is_configured": false, 00:29:50.811 "data_offset": 0, 00:29:50.811 "data_size": 0 00:29:50.811 }, 00:29:50.811 { 00:29:50.811 "name": "BaseBdev2", 00:29:50.811 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:50.811 "is_configured": false, 00:29:50.811 "data_offset": 0, 00:29:50.811 "data_size": 0 00:29:50.811 }, 00:29:50.811 { 00:29:50.811 "name": "BaseBdev3", 00:29:50.811 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:50.811 "is_configured": false, 00:29:50.811 "data_offset": 0, 00:29:50.811 "data_size": 0 00:29:50.811 }, 00:29:50.811 { 00:29:50.811 "name": "BaseBdev4", 00:29:50.811 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:50.811 "is_configured": false, 00:29:50.811 "data_offset": 0, 00:29:50.811 "data_size": 0 00:29:50.811 } 00:29:50.811 ] 00:29:50.811 }' 00:29:50.811 13:49:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:50.811 13:49:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:51.069 13:49:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:29:51.069 13:49:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:51.069 13:49:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:51.069 [2024-11-20 13:49:58.003091] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:29:51.069 [2024-11-20 13:49:58.003133] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:29:51.069 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:51.069 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:29:51.069 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:51.069 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:51.069 [2024-11-20 13:49:58.011075] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:29:51.070 [2024-11-20 13:49:58.011113] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:29:51.070 [2024-11-20 13:49:58.011121] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:29:51.070 [2024-11-20 13:49:58.011129] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:29:51.070 [2024-11-20 13:49:58.011134] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:29:51.070 [2024-11-20 13:49:58.011141] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:29:51.070 [2024-11-20 13:49:58.011146] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:29:51.070 [2024-11-20 13:49:58.011154] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:51.070 [2024-11-20 13:49:58.041392] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:29:51.070 BaseBdev1 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:51.070 [ 00:29:51.070 { 00:29:51.070 "name": "BaseBdev1", 00:29:51.070 "aliases": [ 00:29:51.070 "e92385c1-5256-490f-b742-67a4f8a4cb8c" 00:29:51.070 ], 00:29:51.070 "product_name": "Malloc disk", 00:29:51.070 "block_size": 512, 00:29:51.070 "num_blocks": 65536, 00:29:51.070 "uuid": "e92385c1-5256-490f-b742-67a4f8a4cb8c", 00:29:51.070 "assigned_rate_limits": { 00:29:51.070 "rw_ios_per_sec": 0, 00:29:51.070 "rw_mbytes_per_sec": 0, 00:29:51.070 "r_mbytes_per_sec": 0, 00:29:51.070 "w_mbytes_per_sec": 0 00:29:51.070 }, 00:29:51.070 "claimed": true, 00:29:51.070 "claim_type": "exclusive_write", 00:29:51.070 "zoned": false, 00:29:51.070 "supported_io_types": { 00:29:51.070 "read": true, 00:29:51.070 "write": true, 00:29:51.070 "unmap": true, 00:29:51.070 "flush": true, 00:29:51.070 "reset": true, 00:29:51.070 "nvme_admin": false, 00:29:51.070 "nvme_io": false, 00:29:51.070 "nvme_io_md": false, 00:29:51.070 "write_zeroes": true, 00:29:51.070 "zcopy": true, 00:29:51.070 "get_zone_info": false, 00:29:51.070 "zone_management": false, 00:29:51.070 "zone_append": false, 00:29:51.070 "compare": false, 00:29:51.070 "compare_and_write": false, 00:29:51.070 "abort": true, 00:29:51.070 "seek_hole": false, 00:29:51.070 "seek_data": false, 00:29:51.070 "copy": true, 00:29:51.070 "nvme_iov_md": false 00:29:51.070 }, 00:29:51.070 "memory_domains": [ 00:29:51.070 { 00:29:51.070 "dma_device_id": "system", 00:29:51.070 "dma_device_type": 1 00:29:51.070 }, 00:29:51.070 { 00:29:51.070 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:51.070 "dma_device_type": 2 00:29:51.070 } 00:29:51.070 ], 00:29:51.070 "driver_specific": {} 00:29:51.070 } 00:29:51.070 ] 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:51.070 "name": "Existed_Raid", 00:29:51.070 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:51.070 "strip_size_kb": 0, 00:29:51.070 "state": "configuring", 00:29:51.070 "raid_level": "raid1", 00:29:51.070 "superblock": false, 00:29:51.070 "num_base_bdevs": 4, 00:29:51.070 "num_base_bdevs_discovered": 1, 00:29:51.070 "num_base_bdevs_operational": 4, 00:29:51.070 "base_bdevs_list": [ 00:29:51.070 { 00:29:51.070 "name": "BaseBdev1", 00:29:51.070 "uuid": "e92385c1-5256-490f-b742-67a4f8a4cb8c", 00:29:51.070 "is_configured": true, 00:29:51.070 "data_offset": 0, 00:29:51.070 "data_size": 65536 00:29:51.070 }, 00:29:51.070 { 00:29:51.070 "name": "BaseBdev2", 00:29:51.070 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:51.070 "is_configured": false, 00:29:51.070 "data_offset": 0, 00:29:51.070 "data_size": 0 00:29:51.070 }, 00:29:51.070 { 00:29:51.070 "name": "BaseBdev3", 00:29:51.070 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:51.070 "is_configured": false, 00:29:51.070 "data_offset": 0, 00:29:51.070 "data_size": 0 00:29:51.070 }, 00:29:51.070 { 00:29:51.070 "name": "BaseBdev4", 00:29:51.070 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:51.070 "is_configured": false, 00:29:51.070 "data_offset": 0, 00:29:51.070 "data_size": 0 00:29:51.070 } 00:29:51.070 ] 00:29:51.070 }' 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:51.070 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:51.636 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:29:51.636 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:51.636 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:51.636 [2024-11-20 13:49:58.405519] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:29:51.636 [2024-11-20 13:49:58.405702] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:29:51.636 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:51.637 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:29:51.637 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:51.637 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:51.637 [2024-11-20 13:49:58.413551] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:29:51.637 [2024-11-20 13:49:58.415326] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:29:51.637 [2024-11-20 13:49:58.415437] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:29:51.637 [2024-11-20 13:49:58.415489] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:29:51.637 [2024-11-20 13:49:58.415541] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:29:51.637 [2024-11-20 13:49:58.415597] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:29:51.637 [2024-11-20 13:49:58.415618] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:29:51.637 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:51.637 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:29:51.637 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:29:51.637 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:29:51.637 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:51.637 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:51.637 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:29:51.637 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:29:51.637 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:51.637 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:51.637 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:51.637 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:51.637 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:51.637 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:51.637 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:51.637 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:51.637 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:51.637 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:51.637 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:51.637 "name": "Existed_Raid", 00:29:51.637 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:51.637 "strip_size_kb": 0, 00:29:51.637 "state": "configuring", 00:29:51.637 "raid_level": "raid1", 00:29:51.637 "superblock": false, 00:29:51.637 "num_base_bdevs": 4, 00:29:51.637 "num_base_bdevs_discovered": 1, 00:29:51.637 "num_base_bdevs_operational": 4, 00:29:51.637 "base_bdevs_list": [ 00:29:51.637 { 00:29:51.637 "name": "BaseBdev1", 00:29:51.637 "uuid": "e92385c1-5256-490f-b742-67a4f8a4cb8c", 00:29:51.637 "is_configured": true, 00:29:51.637 "data_offset": 0, 00:29:51.637 "data_size": 65536 00:29:51.637 }, 00:29:51.637 { 00:29:51.637 "name": "BaseBdev2", 00:29:51.637 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:51.637 "is_configured": false, 00:29:51.637 "data_offset": 0, 00:29:51.637 "data_size": 0 00:29:51.637 }, 00:29:51.637 { 00:29:51.637 "name": "BaseBdev3", 00:29:51.637 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:51.637 "is_configured": false, 00:29:51.637 "data_offset": 0, 00:29:51.637 "data_size": 0 00:29:51.637 }, 00:29:51.637 { 00:29:51.637 "name": "BaseBdev4", 00:29:51.637 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:51.637 "is_configured": false, 00:29:51.637 "data_offset": 0, 00:29:51.637 "data_size": 0 00:29:51.637 } 00:29:51.637 ] 00:29:51.637 }' 00:29:51.637 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:51.637 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:51.895 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:29:51.895 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:51.895 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:51.895 [2024-11-20 13:49:58.738378] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:29:51.895 BaseBdev2 00:29:51.895 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:51.895 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:29:51.895 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:29:51.895 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:51.895 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:29:51.895 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:51.895 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:51.895 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:51.895 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:51.895 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:51.895 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:51.896 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:29:51.896 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:51.896 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:51.896 [ 00:29:51.896 { 00:29:51.896 "name": "BaseBdev2", 00:29:51.896 "aliases": [ 00:29:51.896 "4f58f5c7-af3a-4d40-9906-6dcf1ee1d7e8" 00:29:51.896 ], 00:29:51.896 "product_name": "Malloc disk", 00:29:51.896 "block_size": 512, 00:29:51.896 "num_blocks": 65536, 00:29:51.896 "uuid": "4f58f5c7-af3a-4d40-9906-6dcf1ee1d7e8", 00:29:51.896 "assigned_rate_limits": { 00:29:51.896 "rw_ios_per_sec": 0, 00:29:51.896 "rw_mbytes_per_sec": 0, 00:29:51.896 "r_mbytes_per_sec": 0, 00:29:51.896 "w_mbytes_per_sec": 0 00:29:51.896 }, 00:29:51.896 "claimed": true, 00:29:51.896 "claim_type": "exclusive_write", 00:29:51.896 "zoned": false, 00:29:51.896 "supported_io_types": { 00:29:51.896 "read": true, 00:29:51.896 "write": true, 00:29:51.896 "unmap": true, 00:29:51.896 "flush": true, 00:29:51.896 "reset": true, 00:29:51.896 "nvme_admin": false, 00:29:51.896 "nvme_io": false, 00:29:51.896 "nvme_io_md": false, 00:29:51.896 "write_zeroes": true, 00:29:51.896 "zcopy": true, 00:29:51.896 "get_zone_info": false, 00:29:51.896 "zone_management": false, 00:29:51.896 "zone_append": false, 00:29:51.896 "compare": false, 00:29:51.896 "compare_and_write": false, 00:29:51.896 "abort": true, 00:29:51.896 "seek_hole": false, 00:29:51.896 "seek_data": false, 00:29:51.896 "copy": true, 00:29:51.896 "nvme_iov_md": false 00:29:51.896 }, 00:29:51.896 "memory_domains": [ 00:29:51.896 { 00:29:51.896 "dma_device_id": "system", 00:29:51.896 "dma_device_type": 1 00:29:51.896 }, 00:29:51.896 { 00:29:51.896 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:51.896 "dma_device_type": 2 00:29:51.896 } 00:29:51.896 ], 00:29:51.896 "driver_specific": {} 00:29:51.896 } 00:29:51.896 ] 00:29:51.896 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:51.896 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:29:51.896 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:29:51.896 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:29:51.896 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:29:51.896 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:51.896 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:51.896 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:29:51.896 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:29:51.896 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:51.896 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:51.896 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:51.896 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:51.896 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:51.896 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:51.896 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:51.896 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:51.896 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:51.896 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:51.896 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:51.896 "name": "Existed_Raid", 00:29:51.896 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:51.896 "strip_size_kb": 0, 00:29:51.896 "state": "configuring", 00:29:51.896 "raid_level": "raid1", 00:29:51.896 "superblock": false, 00:29:51.896 "num_base_bdevs": 4, 00:29:51.896 "num_base_bdevs_discovered": 2, 00:29:51.896 "num_base_bdevs_operational": 4, 00:29:51.896 "base_bdevs_list": [ 00:29:51.896 { 00:29:51.896 "name": "BaseBdev1", 00:29:51.896 "uuid": "e92385c1-5256-490f-b742-67a4f8a4cb8c", 00:29:51.896 "is_configured": true, 00:29:51.896 "data_offset": 0, 00:29:51.896 "data_size": 65536 00:29:51.896 }, 00:29:51.896 { 00:29:51.896 "name": "BaseBdev2", 00:29:51.896 "uuid": "4f58f5c7-af3a-4d40-9906-6dcf1ee1d7e8", 00:29:51.896 "is_configured": true, 00:29:51.896 "data_offset": 0, 00:29:51.896 "data_size": 65536 00:29:51.896 }, 00:29:51.896 { 00:29:51.896 "name": "BaseBdev3", 00:29:51.896 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:51.896 "is_configured": false, 00:29:51.896 "data_offset": 0, 00:29:51.896 "data_size": 0 00:29:51.896 }, 00:29:51.896 { 00:29:51.896 "name": "BaseBdev4", 00:29:51.896 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:51.896 "is_configured": false, 00:29:51.896 "data_offset": 0, 00:29:51.896 "data_size": 0 00:29:51.896 } 00:29:51.896 ] 00:29:51.896 }' 00:29:51.896 13:49:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:51.896 13:49:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:52.154 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:29:52.154 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:52.154 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:52.154 [2024-11-20 13:49:59.135394] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:29:52.154 BaseBdev3 00:29:52.154 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:52.154 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:29:52.154 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:29:52.154 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:52.154 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:29:52.154 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:52.154 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:52.154 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:52.154 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:52.154 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:52.154 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:52.154 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:29:52.154 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:52.154 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:52.154 [ 00:29:52.154 { 00:29:52.154 "name": "BaseBdev3", 00:29:52.154 "aliases": [ 00:29:52.154 "76b5644f-3366-486f-9b5a-85e7a157d694" 00:29:52.154 ], 00:29:52.154 "product_name": "Malloc disk", 00:29:52.154 "block_size": 512, 00:29:52.154 "num_blocks": 65536, 00:29:52.154 "uuid": "76b5644f-3366-486f-9b5a-85e7a157d694", 00:29:52.154 "assigned_rate_limits": { 00:29:52.154 "rw_ios_per_sec": 0, 00:29:52.154 "rw_mbytes_per_sec": 0, 00:29:52.154 "r_mbytes_per_sec": 0, 00:29:52.154 "w_mbytes_per_sec": 0 00:29:52.154 }, 00:29:52.154 "claimed": true, 00:29:52.154 "claim_type": "exclusive_write", 00:29:52.154 "zoned": false, 00:29:52.154 "supported_io_types": { 00:29:52.154 "read": true, 00:29:52.154 "write": true, 00:29:52.154 "unmap": true, 00:29:52.154 "flush": true, 00:29:52.154 "reset": true, 00:29:52.154 "nvme_admin": false, 00:29:52.154 "nvme_io": false, 00:29:52.154 "nvme_io_md": false, 00:29:52.154 "write_zeroes": true, 00:29:52.154 "zcopy": true, 00:29:52.154 "get_zone_info": false, 00:29:52.154 "zone_management": false, 00:29:52.154 "zone_append": false, 00:29:52.154 "compare": false, 00:29:52.155 "compare_and_write": false, 00:29:52.155 "abort": true, 00:29:52.155 "seek_hole": false, 00:29:52.155 "seek_data": false, 00:29:52.155 "copy": true, 00:29:52.155 "nvme_iov_md": false 00:29:52.155 }, 00:29:52.155 "memory_domains": [ 00:29:52.155 { 00:29:52.155 "dma_device_id": "system", 00:29:52.155 "dma_device_type": 1 00:29:52.155 }, 00:29:52.155 { 00:29:52.155 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:52.155 "dma_device_type": 2 00:29:52.155 } 00:29:52.155 ], 00:29:52.155 "driver_specific": {} 00:29:52.155 } 00:29:52.155 ] 00:29:52.155 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:52.155 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:29:52.155 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:29:52.155 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:29:52.155 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:29:52.155 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:52.155 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:52.155 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:29:52.155 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:29:52.155 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:52.155 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:52.155 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:52.155 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:52.155 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:52.155 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:52.155 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:52.155 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:52.155 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:52.155 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:52.155 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:52.155 "name": "Existed_Raid", 00:29:52.155 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:52.155 "strip_size_kb": 0, 00:29:52.155 "state": "configuring", 00:29:52.155 "raid_level": "raid1", 00:29:52.155 "superblock": false, 00:29:52.155 "num_base_bdevs": 4, 00:29:52.155 "num_base_bdevs_discovered": 3, 00:29:52.155 "num_base_bdevs_operational": 4, 00:29:52.155 "base_bdevs_list": [ 00:29:52.155 { 00:29:52.155 "name": "BaseBdev1", 00:29:52.155 "uuid": "e92385c1-5256-490f-b742-67a4f8a4cb8c", 00:29:52.155 "is_configured": true, 00:29:52.155 "data_offset": 0, 00:29:52.155 "data_size": 65536 00:29:52.155 }, 00:29:52.155 { 00:29:52.155 "name": "BaseBdev2", 00:29:52.155 "uuid": "4f58f5c7-af3a-4d40-9906-6dcf1ee1d7e8", 00:29:52.155 "is_configured": true, 00:29:52.155 "data_offset": 0, 00:29:52.155 "data_size": 65536 00:29:52.155 }, 00:29:52.155 { 00:29:52.155 "name": "BaseBdev3", 00:29:52.155 "uuid": "76b5644f-3366-486f-9b5a-85e7a157d694", 00:29:52.155 "is_configured": true, 00:29:52.155 "data_offset": 0, 00:29:52.155 "data_size": 65536 00:29:52.155 }, 00:29:52.155 { 00:29:52.155 "name": "BaseBdev4", 00:29:52.155 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:52.155 "is_configured": false, 00:29:52.155 "data_offset": 0, 00:29:52.155 "data_size": 0 00:29:52.155 } 00:29:52.155 ] 00:29:52.155 }' 00:29:52.155 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:52.155 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:52.720 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:29:52.720 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:52.720 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:52.720 [2024-11-20 13:49:59.528453] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:29:52.720 [2024-11-20 13:49:59.528624] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:29:52.720 [2024-11-20 13:49:59.528653] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:29:52.720 [2024-11-20 13:49:59.528990] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:29:52.720 [2024-11-20 13:49:59.529215] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:29:52.720 [2024-11-20 13:49:59.529295] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:29:52.720 BaseBdev4 00:29:52.720 [2024-11-20 13:49:59.529620] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:52.720 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:52.720 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:29:52.720 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:29:52.720 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:52.720 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:29:52.720 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:52.720 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:52.720 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:52.720 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:52.720 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:52.720 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:52.720 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:29:52.720 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:52.720 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:52.720 [ 00:29:52.720 { 00:29:52.720 "name": "BaseBdev4", 00:29:52.720 "aliases": [ 00:29:52.720 "6ae75838-9319-4d02-be6c-83bb1264f275" 00:29:52.720 ], 00:29:52.720 "product_name": "Malloc disk", 00:29:52.720 "block_size": 512, 00:29:52.720 "num_blocks": 65536, 00:29:52.720 "uuid": "6ae75838-9319-4d02-be6c-83bb1264f275", 00:29:52.720 "assigned_rate_limits": { 00:29:52.720 "rw_ios_per_sec": 0, 00:29:52.720 "rw_mbytes_per_sec": 0, 00:29:52.720 "r_mbytes_per_sec": 0, 00:29:52.720 "w_mbytes_per_sec": 0 00:29:52.720 }, 00:29:52.720 "claimed": true, 00:29:52.720 "claim_type": "exclusive_write", 00:29:52.720 "zoned": false, 00:29:52.720 "supported_io_types": { 00:29:52.720 "read": true, 00:29:52.720 "write": true, 00:29:52.720 "unmap": true, 00:29:52.720 "flush": true, 00:29:52.720 "reset": true, 00:29:52.720 "nvme_admin": false, 00:29:52.720 "nvme_io": false, 00:29:52.720 "nvme_io_md": false, 00:29:52.720 "write_zeroes": true, 00:29:52.720 "zcopy": true, 00:29:52.720 "get_zone_info": false, 00:29:52.720 "zone_management": false, 00:29:52.720 "zone_append": false, 00:29:52.720 "compare": false, 00:29:52.721 "compare_and_write": false, 00:29:52.721 "abort": true, 00:29:52.721 "seek_hole": false, 00:29:52.721 "seek_data": false, 00:29:52.721 "copy": true, 00:29:52.721 "nvme_iov_md": false 00:29:52.721 }, 00:29:52.721 "memory_domains": [ 00:29:52.721 { 00:29:52.721 "dma_device_id": "system", 00:29:52.721 "dma_device_type": 1 00:29:52.721 }, 00:29:52.721 { 00:29:52.721 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:52.721 "dma_device_type": 2 00:29:52.721 } 00:29:52.721 ], 00:29:52.721 "driver_specific": {} 00:29:52.721 } 00:29:52.721 ] 00:29:52.721 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:52.721 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:29:52.721 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:29:52.721 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:29:52.721 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:29:52.721 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:52.721 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:52.721 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:29:52.721 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:29:52.721 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:52.721 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:52.721 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:52.721 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:52.721 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:52.721 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:52.721 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:52.721 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:52.721 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:52.721 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:52.721 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:52.721 "name": "Existed_Raid", 00:29:52.721 "uuid": "634d0e71-1421-4c02-8660-cf836aae4315", 00:29:52.721 "strip_size_kb": 0, 00:29:52.721 "state": "online", 00:29:52.721 "raid_level": "raid1", 00:29:52.721 "superblock": false, 00:29:52.721 "num_base_bdevs": 4, 00:29:52.721 "num_base_bdevs_discovered": 4, 00:29:52.721 "num_base_bdevs_operational": 4, 00:29:52.721 "base_bdevs_list": [ 00:29:52.721 { 00:29:52.721 "name": "BaseBdev1", 00:29:52.721 "uuid": "e92385c1-5256-490f-b742-67a4f8a4cb8c", 00:29:52.721 "is_configured": true, 00:29:52.721 "data_offset": 0, 00:29:52.721 "data_size": 65536 00:29:52.721 }, 00:29:52.721 { 00:29:52.721 "name": "BaseBdev2", 00:29:52.721 "uuid": "4f58f5c7-af3a-4d40-9906-6dcf1ee1d7e8", 00:29:52.721 "is_configured": true, 00:29:52.721 "data_offset": 0, 00:29:52.721 "data_size": 65536 00:29:52.721 }, 00:29:52.721 { 00:29:52.721 "name": "BaseBdev3", 00:29:52.721 "uuid": "76b5644f-3366-486f-9b5a-85e7a157d694", 00:29:52.721 "is_configured": true, 00:29:52.721 "data_offset": 0, 00:29:52.721 "data_size": 65536 00:29:52.721 }, 00:29:52.721 { 00:29:52.721 "name": "BaseBdev4", 00:29:52.721 "uuid": "6ae75838-9319-4d02-be6c-83bb1264f275", 00:29:52.721 "is_configured": true, 00:29:52.721 "data_offset": 0, 00:29:52.721 "data_size": 65536 00:29:52.721 } 00:29:52.721 ] 00:29:52.721 }' 00:29:52.721 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:52.721 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:52.979 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:29:52.979 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:29:52.979 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:29:52.979 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:29:52.979 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:29:52.979 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:29:52.979 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:29:52.979 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:29:52.979 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:52.979 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:52.979 [2024-11-20 13:49:59.884905] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:52.979 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:52.979 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:52.979 "name": "Existed_Raid", 00:29:52.979 "aliases": [ 00:29:52.979 "634d0e71-1421-4c02-8660-cf836aae4315" 00:29:52.979 ], 00:29:52.979 "product_name": "Raid Volume", 00:29:52.979 "block_size": 512, 00:29:52.980 "num_blocks": 65536, 00:29:52.980 "uuid": "634d0e71-1421-4c02-8660-cf836aae4315", 00:29:52.980 "assigned_rate_limits": { 00:29:52.980 "rw_ios_per_sec": 0, 00:29:52.980 "rw_mbytes_per_sec": 0, 00:29:52.980 "r_mbytes_per_sec": 0, 00:29:52.980 "w_mbytes_per_sec": 0 00:29:52.980 }, 00:29:52.980 "claimed": false, 00:29:52.980 "zoned": false, 00:29:52.980 "supported_io_types": { 00:29:52.980 "read": true, 00:29:52.980 "write": true, 00:29:52.980 "unmap": false, 00:29:52.980 "flush": false, 00:29:52.980 "reset": true, 00:29:52.980 "nvme_admin": false, 00:29:52.980 "nvme_io": false, 00:29:52.980 "nvme_io_md": false, 00:29:52.980 "write_zeroes": true, 00:29:52.980 "zcopy": false, 00:29:52.980 "get_zone_info": false, 00:29:52.980 "zone_management": false, 00:29:52.980 "zone_append": false, 00:29:52.980 "compare": false, 00:29:52.980 "compare_and_write": false, 00:29:52.980 "abort": false, 00:29:52.980 "seek_hole": false, 00:29:52.980 "seek_data": false, 00:29:52.980 "copy": false, 00:29:52.980 "nvme_iov_md": false 00:29:52.980 }, 00:29:52.980 "memory_domains": [ 00:29:52.980 { 00:29:52.980 "dma_device_id": "system", 00:29:52.980 "dma_device_type": 1 00:29:52.980 }, 00:29:52.980 { 00:29:52.980 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:52.980 "dma_device_type": 2 00:29:52.980 }, 00:29:52.980 { 00:29:52.980 "dma_device_id": "system", 00:29:52.980 "dma_device_type": 1 00:29:52.980 }, 00:29:52.980 { 00:29:52.980 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:52.980 "dma_device_type": 2 00:29:52.980 }, 00:29:52.980 { 00:29:52.980 "dma_device_id": "system", 00:29:52.980 "dma_device_type": 1 00:29:52.980 }, 00:29:52.980 { 00:29:52.980 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:52.980 "dma_device_type": 2 00:29:52.980 }, 00:29:52.980 { 00:29:52.980 "dma_device_id": "system", 00:29:52.980 "dma_device_type": 1 00:29:52.980 }, 00:29:52.980 { 00:29:52.980 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:52.980 "dma_device_type": 2 00:29:52.980 } 00:29:52.980 ], 00:29:52.980 "driver_specific": { 00:29:52.980 "raid": { 00:29:52.980 "uuid": "634d0e71-1421-4c02-8660-cf836aae4315", 00:29:52.980 "strip_size_kb": 0, 00:29:52.980 "state": "online", 00:29:52.980 "raid_level": "raid1", 00:29:52.980 "superblock": false, 00:29:52.980 "num_base_bdevs": 4, 00:29:52.980 "num_base_bdevs_discovered": 4, 00:29:52.980 "num_base_bdevs_operational": 4, 00:29:52.980 "base_bdevs_list": [ 00:29:52.980 { 00:29:52.980 "name": "BaseBdev1", 00:29:52.980 "uuid": "e92385c1-5256-490f-b742-67a4f8a4cb8c", 00:29:52.980 "is_configured": true, 00:29:52.980 "data_offset": 0, 00:29:52.980 "data_size": 65536 00:29:52.980 }, 00:29:52.980 { 00:29:52.980 "name": "BaseBdev2", 00:29:52.980 "uuid": "4f58f5c7-af3a-4d40-9906-6dcf1ee1d7e8", 00:29:52.980 "is_configured": true, 00:29:52.980 "data_offset": 0, 00:29:52.980 "data_size": 65536 00:29:52.980 }, 00:29:52.980 { 00:29:52.980 "name": "BaseBdev3", 00:29:52.980 "uuid": "76b5644f-3366-486f-9b5a-85e7a157d694", 00:29:52.980 "is_configured": true, 00:29:52.980 "data_offset": 0, 00:29:52.980 "data_size": 65536 00:29:52.980 }, 00:29:52.980 { 00:29:52.980 "name": "BaseBdev4", 00:29:52.980 "uuid": "6ae75838-9319-4d02-be6c-83bb1264f275", 00:29:52.980 "is_configured": true, 00:29:52.980 "data_offset": 0, 00:29:52.980 "data_size": 65536 00:29:52.980 } 00:29:52.980 ] 00:29:52.980 } 00:29:52.980 } 00:29:52.980 }' 00:29:52.980 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:29:52.980 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:29:52.980 BaseBdev2 00:29:52.980 BaseBdev3 00:29:52.980 BaseBdev4' 00:29:52.980 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:52.980 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:29:52.980 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:52.980 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:29:52.980 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:52.980 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:52.980 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:52.980 13:49:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:52.980 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:52.980 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:52.980 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:52.980 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:29:52.980 13:49:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:52.980 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:52.980 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:52.980 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:52.980 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:52.980 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:52.980 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:52.980 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:29:52.980 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:52.980 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:52.980 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:53.238 [2024-11-20 13:50:00.092690] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:53.238 "name": "Existed_Raid", 00:29:53.238 "uuid": "634d0e71-1421-4c02-8660-cf836aae4315", 00:29:53.238 "strip_size_kb": 0, 00:29:53.238 "state": "online", 00:29:53.238 "raid_level": "raid1", 00:29:53.238 "superblock": false, 00:29:53.238 "num_base_bdevs": 4, 00:29:53.238 "num_base_bdevs_discovered": 3, 00:29:53.238 "num_base_bdevs_operational": 3, 00:29:53.238 "base_bdevs_list": [ 00:29:53.238 { 00:29:53.238 "name": null, 00:29:53.238 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:53.238 "is_configured": false, 00:29:53.238 "data_offset": 0, 00:29:53.238 "data_size": 65536 00:29:53.238 }, 00:29:53.238 { 00:29:53.238 "name": "BaseBdev2", 00:29:53.238 "uuid": "4f58f5c7-af3a-4d40-9906-6dcf1ee1d7e8", 00:29:53.238 "is_configured": true, 00:29:53.238 "data_offset": 0, 00:29:53.238 "data_size": 65536 00:29:53.238 }, 00:29:53.238 { 00:29:53.238 "name": "BaseBdev3", 00:29:53.238 "uuid": "76b5644f-3366-486f-9b5a-85e7a157d694", 00:29:53.238 "is_configured": true, 00:29:53.238 "data_offset": 0, 00:29:53.238 "data_size": 65536 00:29:53.238 }, 00:29:53.238 { 00:29:53.238 "name": "BaseBdev4", 00:29:53.238 "uuid": "6ae75838-9319-4d02-be6c-83bb1264f275", 00:29:53.238 "is_configured": true, 00:29:53.238 "data_offset": 0, 00:29:53.238 "data_size": 65536 00:29:53.238 } 00:29:53.238 ] 00:29:53.238 }' 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:53.238 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:53.496 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:29:53.496 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:29:53.496 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:29:53.496 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:53.496 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:53.496 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:53.496 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:53.496 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:29:53.496 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:29:53.496 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:29:53.496 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:53.496 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:53.496 [2024-11-20 13:50:00.491790] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:29:53.496 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:53.496 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:29:53.496 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:29:53.496 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:29:53.496 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:53.496 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:53.496 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:53.754 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:53.754 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:29:53.754 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:29:53.754 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:29:53.754 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:53.754 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:53.754 [2024-11-20 13:50:00.581181] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:29:53.754 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:53.754 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:29:53.754 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:29:53.754 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:53.754 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:53.754 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:53.754 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:29:53.754 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:53.754 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:29:53.754 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:29:53.754 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:29:53.754 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:53.754 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:53.754 [2024-11-20 13:50:00.671228] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:29:53.754 [2024-11-20 13:50:00.671322] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:29:53.754 [2024-11-20 13:50:00.721694] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:53.754 [2024-11-20 13:50:00.721745] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:29:53.754 [2024-11-20 13:50:00.721757] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:29:53.754 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:53.754 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:29:53.754 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:29:53.754 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:53.755 BaseBdev2 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:53.755 [ 00:29:53.755 { 00:29:53.755 "name": "BaseBdev2", 00:29:53.755 "aliases": [ 00:29:53.755 "45e48512-b083-460d-bd93-f1cae8ce40cf" 00:29:53.755 ], 00:29:53.755 "product_name": "Malloc disk", 00:29:53.755 "block_size": 512, 00:29:53.755 "num_blocks": 65536, 00:29:53.755 "uuid": "45e48512-b083-460d-bd93-f1cae8ce40cf", 00:29:53.755 "assigned_rate_limits": { 00:29:53.755 "rw_ios_per_sec": 0, 00:29:53.755 "rw_mbytes_per_sec": 0, 00:29:53.755 "r_mbytes_per_sec": 0, 00:29:53.755 "w_mbytes_per_sec": 0 00:29:53.755 }, 00:29:53.755 "claimed": false, 00:29:53.755 "zoned": false, 00:29:53.755 "supported_io_types": { 00:29:53.755 "read": true, 00:29:53.755 "write": true, 00:29:53.755 "unmap": true, 00:29:53.755 "flush": true, 00:29:53.755 "reset": true, 00:29:53.755 "nvme_admin": false, 00:29:53.755 "nvme_io": false, 00:29:53.755 "nvme_io_md": false, 00:29:53.755 "write_zeroes": true, 00:29:53.755 "zcopy": true, 00:29:53.755 "get_zone_info": false, 00:29:53.755 "zone_management": false, 00:29:53.755 "zone_append": false, 00:29:53.755 "compare": false, 00:29:53.755 "compare_and_write": false, 00:29:53.755 "abort": true, 00:29:53.755 "seek_hole": false, 00:29:53.755 "seek_data": false, 00:29:53.755 "copy": true, 00:29:53.755 "nvme_iov_md": false 00:29:53.755 }, 00:29:53.755 "memory_domains": [ 00:29:53.755 { 00:29:53.755 "dma_device_id": "system", 00:29:53.755 "dma_device_type": 1 00:29:53.755 }, 00:29:53.755 { 00:29:53.755 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:53.755 "dma_device_type": 2 00:29:53.755 } 00:29:53.755 ], 00:29:53.755 "driver_specific": {} 00:29:53.755 } 00:29:53.755 ] 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:53.755 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:54.013 BaseBdev3 00:29:54.013 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:54.013 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:29:54.013 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:29:54.013 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:54.013 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:29:54.013 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:54.013 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:54.013 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:54.013 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:54.013 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:54.013 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:54.013 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:29:54.013 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:54.013 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:54.013 [ 00:29:54.013 { 00:29:54.013 "name": "BaseBdev3", 00:29:54.013 "aliases": [ 00:29:54.013 "564616e3-0aae-4eb1-9509-df8aa672175d" 00:29:54.013 ], 00:29:54.013 "product_name": "Malloc disk", 00:29:54.013 "block_size": 512, 00:29:54.013 "num_blocks": 65536, 00:29:54.013 "uuid": "564616e3-0aae-4eb1-9509-df8aa672175d", 00:29:54.013 "assigned_rate_limits": { 00:29:54.013 "rw_ios_per_sec": 0, 00:29:54.013 "rw_mbytes_per_sec": 0, 00:29:54.013 "r_mbytes_per_sec": 0, 00:29:54.013 "w_mbytes_per_sec": 0 00:29:54.013 }, 00:29:54.013 "claimed": false, 00:29:54.013 "zoned": false, 00:29:54.013 "supported_io_types": { 00:29:54.013 "read": true, 00:29:54.013 "write": true, 00:29:54.013 "unmap": true, 00:29:54.013 "flush": true, 00:29:54.013 "reset": true, 00:29:54.013 "nvme_admin": false, 00:29:54.013 "nvme_io": false, 00:29:54.013 "nvme_io_md": false, 00:29:54.013 "write_zeroes": true, 00:29:54.013 "zcopy": true, 00:29:54.013 "get_zone_info": false, 00:29:54.013 "zone_management": false, 00:29:54.013 "zone_append": false, 00:29:54.013 "compare": false, 00:29:54.013 "compare_and_write": false, 00:29:54.013 "abort": true, 00:29:54.013 "seek_hole": false, 00:29:54.013 "seek_data": false, 00:29:54.013 "copy": true, 00:29:54.013 "nvme_iov_md": false 00:29:54.013 }, 00:29:54.013 "memory_domains": [ 00:29:54.013 { 00:29:54.013 "dma_device_id": "system", 00:29:54.013 "dma_device_type": 1 00:29:54.013 }, 00:29:54.013 { 00:29:54.013 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:54.013 "dma_device_type": 2 00:29:54.013 } 00:29:54.013 ], 00:29:54.013 "driver_specific": {} 00:29:54.013 } 00:29:54.013 ] 00:29:54.013 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:54.013 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:29:54.013 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:54.014 BaseBdev4 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:54.014 [ 00:29:54.014 { 00:29:54.014 "name": "BaseBdev4", 00:29:54.014 "aliases": [ 00:29:54.014 "4f510a7a-c2f6-4f78-99d3-d27ef3a42e7b" 00:29:54.014 ], 00:29:54.014 "product_name": "Malloc disk", 00:29:54.014 "block_size": 512, 00:29:54.014 "num_blocks": 65536, 00:29:54.014 "uuid": "4f510a7a-c2f6-4f78-99d3-d27ef3a42e7b", 00:29:54.014 "assigned_rate_limits": { 00:29:54.014 "rw_ios_per_sec": 0, 00:29:54.014 "rw_mbytes_per_sec": 0, 00:29:54.014 "r_mbytes_per_sec": 0, 00:29:54.014 "w_mbytes_per_sec": 0 00:29:54.014 }, 00:29:54.014 "claimed": false, 00:29:54.014 "zoned": false, 00:29:54.014 "supported_io_types": { 00:29:54.014 "read": true, 00:29:54.014 "write": true, 00:29:54.014 "unmap": true, 00:29:54.014 "flush": true, 00:29:54.014 "reset": true, 00:29:54.014 "nvme_admin": false, 00:29:54.014 "nvme_io": false, 00:29:54.014 "nvme_io_md": false, 00:29:54.014 "write_zeroes": true, 00:29:54.014 "zcopy": true, 00:29:54.014 "get_zone_info": false, 00:29:54.014 "zone_management": false, 00:29:54.014 "zone_append": false, 00:29:54.014 "compare": false, 00:29:54.014 "compare_and_write": false, 00:29:54.014 "abort": true, 00:29:54.014 "seek_hole": false, 00:29:54.014 "seek_data": false, 00:29:54.014 "copy": true, 00:29:54.014 "nvme_iov_md": false 00:29:54.014 }, 00:29:54.014 "memory_domains": [ 00:29:54.014 { 00:29:54.014 "dma_device_id": "system", 00:29:54.014 "dma_device_type": 1 00:29:54.014 }, 00:29:54.014 { 00:29:54.014 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:54.014 "dma_device_type": 2 00:29:54.014 } 00:29:54.014 ], 00:29:54.014 "driver_specific": {} 00:29:54.014 } 00:29:54.014 ] 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:54.014 [2024-11-20 13:50:00.912077] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:29:54.014 [2024-11-20 13:50:00.912238] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:29:54.014 [2024-11-20 13:50:00.912300] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:29:54.014 [2024-11-20 13:50:00.914192] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:29:54.014 [2024-11-20 13:50:00.914310] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:54.014 "name": "Existed_Raid", 00:29:54.014 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:54.014 "strip_size_kb": 0, 00:29:54.014 "state": "configuring", 00:29:54.014 "raid_level": "raid1", 00:29:54.014 "superblock": false, 00:29:54.014 "num_base_bdevs": 4, 00:29:54.014 "num_base_bdevs_discovered": 3, 00:29:54.014 "num_base_bdevs_operational": 4, 00:29:54.014 "base_bdevs_list": [ 00:29:54.014 { 00:29:54.014 "name": "BaseBdev1", 00:29:54.014 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:54.014 "is_configured": false, 00:29:54.014 "data_offset": 0, 00:29:54.014 "data_size": 0 00:29:54.014 }, 00:29:54.014 { 00:29:54.014 "name": "BaseBdev2", 00:29:54.014 "uuid": "45e48512-b083-460d-bd93-f1cae8ce40cf", 00:29:54.014 "is_configured": true, 00:29:54.014 "data_offset": 0, 00:29:54.014 "data_size": 65536 00:29:54.014 }, 00:29:54.014 { 00:29:54.014 "name": "BaseBdev3", 00:29:54.014 "uuid": "564616e3-0aae-4eb1-9509-df8aa672175d", 00:29:54.014 "is_configured": true, 00:29:54.014 "data_offset": 0, 00:29:54.014 "data_size": 65536 00:29:54.014 }, 00:29:54.014 { 00:29:54.014 "name": "BaseBdev4", 00:29:54.014 "uuid": "4f510a7a-c2f6-4f78-99d3-d27ef3a42e7b", 00:29:54.014 "is_configured": true, 00:29:54.014 "data_offset": 0, 00:29:54.014 "data_size": 65536 00:29:54.014 } 00:29:54.014 ] 00:29:54.014 }' 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:54.014 13:50:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:54.272 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:29:54.272 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:54.272 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:54.272 [2024-11-20 13:50:01.228163] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:29:54.272 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:54.272 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:29:54.272 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:54.272 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:54.272 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:29:54.272 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:29:54.272 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:54.272 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:54.272 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:54.272 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:54.272 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:54.272 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:54.272 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:54.272 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:54.272 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:54.273 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:54.273 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:54.273 "name": "Existed_Raid", 00:29:54.273 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:54.273 "strip_size_kb": 0, 00:29:54.273 "state": "configuring", 00:29:54.273 "raid_level": "raid1", 00:29:54.273 "superblock": false, 00:29:54.273 "num_base_bdevs": 4, 00:29:54.273 "num_base_bdevs_discovered": 2, 00:29:54.273 "num_base_bdevs_operational": 4, 00:29:54.273 "base_bdevs_list": [ 00:29:54.273 { 00:29:54.273 "name": "BaseBdev1", 00:29:54.273 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:54.273 "is_configured": false, 00:29:54.273 "data_offset": 0, 00:29:54.273 "data_size": 0 00:29:54.273 }, 00:29:54.273 { 00:29:54.273 "name": null, 00:29:54.273 "uuid": "45e48512-b083-460d-bd93-f1cae8ce40cf", 00:29:54.273 "is_configured": false, 00:29:54.273 "data_offset": 0, 00:29:54.273 "data_size": 65536 00:29:54.273 }, 00:29:54.273 { 00:29:54.273 "name": "BaseBdev3", 00:29:54.273 "uuid": "564616e3-0aae-4eb1-9509-df8aa672175d", 00:29:54.273 "is_configured": true, 00:29:54.273 "data_offset": 0, 00:29:54.273 "data_size": 65536 00:29:54.273 }, 00:29:54.273 { 00:29:54.273 "name": "BaseBdev4", 00:29:54.273 "uuid": "4f510a7a-c2f6-4f78-99d3-d27ef3a42e7b", 00:29:54.273 "is_configured": true, 00:29:54.273 "data_offset": 0, 00:29:54.273 "data_size": 65536 00:29:54.273 } 00:29:54.273 ] 00:29:54.273 }' 00:29:54.273 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:54.273 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:54.838 [2024-11-20 13:50:01.648887] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:29:54.838 BaseBdev1 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:54.838 [ 00:29:54.838 { 00:29:54.838 "name": "BaseBdev1", 00:29:54.838 "aliases": [ 00:29:54.838 "bcfd4a30-5d71-4e62-8e02-cc2bebbed34b" 00:29:54.838 ], 00:29:54.838 "product_name": "Malloc disk", 00:29:54.838 "block_size": 512, 00:29:54.838 "num_blocks": 65536, 00:29:54.838 "uuid": "bcfd4a30-5d71-4e62-8e02-cc2bebbed34b", 00:29:54.838 "assigned_rate_limits": { 00:29:54.838 "rw_ios_per_sec": 0, 00:29:54.838 "rw_mbytes_per_sec": 0, 00:29:54.838 "r_mbytes_per_sec": 0, 00:29:54.838 "w_mbytes_per_sec": 0 00:29:54.838 }, 00:29:54.838 "claimed": true, 00:29:54.838 "claim_type": "exclusive_write", 00:29:54.838 "zoned": false, 00:29:54.838 "supported_io_types": { 00:29:54.838 "read": true, 00:29:54.838 "write": true, 00:29:54.838 "unmap": true, 00:29:54.838 "flush": true, 00:29:54.838 "reset": true, 00:29:54.838 "nvme_admin": false, 00:29:54.838 "nvme_io": false, 00:29:54.838 "nvme_io_md": false, 00:29:54.838 "write_zeroes": true, 00:29:54.838 "zcopy": true, 00:29:54.838 "get_zone_info": false, 00:29:54.838 "zone_management": false, 00:29:54.838 "zone_append": false, 00:29:54.838 "compare": false, 00:29:54.838 "compare_and_write": false, 00:29:54.838 "abort": true, 00:29:54.838 "seek_hole": false, 00:29:54.838 "seek_data": false, 00:29:54.838 "copy": true, 00:29:54.838 "nvme_iov_md": false 00:29:54.838 }, 00:29:54.838 "memory_domains": [ 00:29:54.838 { 00:29:54.838 "dma_device_id": "system", 00:29:54.838 "dma_device_type": 1 00:29:54.838 }, 00:29:54.838 { 00:29:54.838 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:54.838 "dma_device_type": 2 00:29:54.838 } 00:29:54.838 ], 00:29:54.838 "driver_specific": {} 00:29:54.838 } 00:29:54.838 ] 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:54.838 "name": "Existed_Raid", 00:29:54.838 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:54.838 "strip_size_kb": 0, 00:29:54.838 "state": "configuring", 00:29:54.838 "raid_level": "raid1", 00:29:54.838 "superblock": false, 00:29:54.838 "num_base_bdevs": 4, 00:29:54.838 "num_base_bdevs_discovered": 3, 00:29:54.838 "num_base_bdevs_operational": 4, 00:29:54.838 "base_bdevs_list": [ 00:29:54.838 { 00:29:54.838 "name": "BaseBdev1", 00:29:54.838 "uuid": "bcfd4a30-5d71-4e62-8e02-cc2bebbed34b", 00:29:54.838 "is_configured": true, 00:29:54.838 "data_offset": 0, 00:29:54.838 "data_size": 65536 00:29:54.838 }, 00:29:54.838 { 00:29:54.838 "name": null, 00:29:54.838 "uuid": "45e48512-b083-460d-bd93-f1cae8ce40cf", 00:29:54.838 "is_configured": false, 00:29:54.838 "data_offset": 0, 00:29:54.838 "data_size": 65536 00:29:54.838 }, 00:29:54.838 { 00:29:54.838 "name": "BaseBdev3", 00:29:54.838 "uuid": "564616e3-0aae-4eb1-9509-df8aa672175d", 00:29:54.838 "is_configured": true, 00:29:54.838 "data_offset": 0, 00:29:54.838 "data_size": 65536 00:29:54.838 }, 00:29:54.838 { 00:29:54.838 "name": "BaseBdev4", 00:29:54.838 "uuid": "4f510a7a-c2f6-4f78-99d3-d27ef3a42e7b", 00:29:54.838 "is_configured": true, 00:29:54.838 "data_offset": 0, 00:29:54.838 "data_size": 65536 00:29:54.838 } 00:29:54.838 ] 00:29:54.838 }' 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:54.838 13:50:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:55.096 [2024-11-20 13:50:02.069056] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:55.096 "name": "Existed_Raid", 00:29:55.096 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:55.096 "strip_size_kb": 0, 00:29:55.096 "state": "configuring", 00:29:55.096 "raid_level": "raid1", 00:29:55.096 "superblock": false, 00:29:55.096 "num_base_bdevs": 4, 00:29:55.096 "num_base_bdevs_discovered": 2, 00:29:55.096 "num_base_bdevs_operational": 4, 00:29:55.096 "base_bdevs_list": [ 00:29:55.096 { 00:29:55.096 "name": "BaseBdev1", 00:29:55.096 "uuid": "bcfd4a30-5d71-4e62-8e02-cc2bebbed34b", 00:29:55.096 "is_configured": true, 00:29:55.096 "data_offset": 0, 00:29:55.096 "data_size": 65536 00:29:55.096 }, 00:29:55.096 { 00:29:55.096 "name": null, 00:29:55.096 "uuid": "45e48512-b083-460d-bd93-f1cae8ce40cf", 00:29:55.096 "is_configured": false, 00:29:55.096 "data_offset": 0, 00:29:55.096 "data_size": 65536 00:29:55.096 }, 00:29:55.096 { 00:29:55.096 "name": null, 00:29:55.096 "uuid": "564616e3-0aae-4eb1-9509-df8aa672175d", 00:29:55.096 "is_configured": false, 00:29:55.096 "data_offset": 0, 00:29:55.096 "data_size": 65536 00:29:55.096 }, 00:29:55.096 { 00:29:55.096 "name": "BaseBdev4", 00:29:55.096 "uuid": "4f510a7a-c2f6-4f78-99d3-d27ef3a42e7b", 00:29:55.096 "is_configured": true, 00:29:55.096 "data_offset": 0, 00:29:55.096 "data_size": 65536 00:29:55.096 } 00:29:55.096 ] 00:29:55.096 }' 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:55.096 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:55.354 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:29:55.354 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:55.354 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:55.354 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:55.612 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:55.612 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:29:55.612 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:29:55.612 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:55.612 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:55.612 [2024-11-20 13:50:02.433101] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:29:55.612 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:55.612 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:29:55.612 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:55.612 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:55.612 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:29:55.612 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:29:55.612 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:55.612 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:55.612 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:55.612 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:55.612 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:55.612 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:55.612 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:55.612 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:55.612 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:55.612 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:55.612 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:55.612 "name": "Existed_Raid", 00:29:55.612 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:55.612 "strip_size_kb": 0, 00:29:55.612 "state": "configuring", 00:29:55.612 "raid_level": "raid1", 00:29:55.612 "superblock": false, 00:29:55.612 "num_base_bdevs": 4, 00:29:55.612 "num_base_bdevs_discovered": 3, 00:29:55.612 "num_base_bdevs_operational": 4, 00:29:55.612 "base_bdevs_list": [ 00:29:55.612 { 00:29:55.612 "name": "BaseBdev1", 00:29:55.612 "uuid": "bcfd4a30-5d71-4e62-8e02-cc2bebbed34b", 00:29:55.612 "is_configured": true, 00:29:55.612 "data_offset": 0, 00:29:55.612 "data_size": 65536 00:29:55.612 }, 00:29:55.612 { 00:29:55.612 "name": null, 00:29:55.612 "uuid": "45e48512-b083-460d-bd93-f1cae8ce40cf", 00:29:55.612 "is_configured": false, 00:29:55.612 "data_offset": 0, 00:29:55.612 "data_size": 65536 00:29:55.612 }, 00:29:55.612 { 00:29:55.612 "name": "BaseBdev3", 00:29:55.612 "uuid": "564616e3-0aae-4eb1-9509-df8aa672175d", 00:29:55.612 "is_configured": true, 00:29:55.612 "data_offset": 0, 00:29:55.612 "data_size": 65536 00:29:55.612 }, 00:29:55.612 { 00:29:55.612 "name": "BaseBdev4", 00:29:55.612 "uuid": "4f510a7a-c2f6-4f78-99d3-d27ef3a42e7b", 00:29:55.612 "is_configured": true, 00:29:55.612 "data_offset": 0, 00:29:55.612 "data_size": 65536 00:29:55.612 } 00:29:55.612 ] 00:29:55.612 }' 00:29:55.612 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:55.612 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:55.871 [2024-11-20 13:50:02.793239] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:55.871 "name": "Existed_Raid", 00:29:55.871 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:55.871 "strip_size_kb": 0, 00:29:55.871 "state": "configuring", 00:29:55.871 "raid_level": "raid1", 00:29:55.871 "superblock": false, 00:29:55.871 "num_base_bdevs": 4, 00:29:55.871 "num_base_bdevs_discovered": 2, 00:29:55.871 "num_base_bdevs_operational": 4, 00:29:55.871 "base_bdevs_list": [ 00:29:55.871 { 00:29:55.871 "name": null, 00:29:55.871 "uuid": "bcfd4a30-5d71-4e62-8e02-cc2bebbed34b", 00:29:55.871 "is_configured": false, 00:29:55.871 "data_offset": 0, 00:29:55.871 "data_size": 65536 00:29:55.871 }, 00:29:55.871 { 00:29:55.871 "name": null, 00:29:55.871 "uuid": "45e48512-b083-460d-bd93-f1cae8ce40cf", 00:29:55.871 "is_configured": false, 00:29:55.871 "data_offset": 0, 00:29:55.871 "data_size": 65536 00:29:55.871 }, 00:29:55.871 { 00:29:55.871 "name": "BaseBdev3", 00:29:55.871 "uuid": "564616e3-0aae-4eb1-9509-df8aa672175d", 00:29:55.871 "is_configured": true, 00:29:55.871 "data_offset": 0, 00:29:55.871 "data_size": 65536 00:29:55.871 }, 00:29:55.871 { 00:29:55.871 "name": "BaseBdev4", 00:29:55.871 "uuid": "4f510a7a-c2f6-4f78-99d3-d27ef3a42e7b", 00:29:55.871 "is_configured": true, 00:29:55.871 "data_offset": 0, 00:29:55.871 "data_size": 65536 00:29:55.871 } 00:29:55.871 ] 00:29:55.871 }' 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:55.871 13:50:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:56.508 [2024-11-20 13:50:03.231201] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:56.508 "name": "Existed_Raid", 00:29:56.508 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:56.508 "strip_size_kb": 0, 00:29:56.508 "state": "configuring", 00:29:56.508 "raid_level": "raid1", 00:29:56.508 "superblock": false, 00:29:56.508 "num_base_bdevs": 4, 00:29:56.508 "num_base_bdevs_discovered": 3, 00:29:56.508 "num_base_bdevs_operational": 4, 00:29:56.508 "base_bdevs_list": [ 00:29:56.508 { 00:29:56.508 "name": null, 00:29:56.508 "uuid": "bcfd4a30-5d71-4e62-8e02-cc2bebbed34b", 00:29:56.508 "is_configured": false, 00:29:56.508 "data_offset": 0, 00:29:56.508 "data_size": 65536 00:29:56.508 }, 00:29:56.508 { 00:29:56.508 "name": "BaseBdev2", 00:29:56.508 "uuid": "45e48512-b083-460d-bd93-f1cae8ce40cf", 00:29:56.508 "is_configured": true, 00:29:56.508 "data_offset": 0, 00:29:56.508 "data_size": 65536 00:29:56.508 }, 00:29:56.508 { 00:29:56.508 "name": "BaseBdev3", 00:29:56.508 "uuid": "564616e3-0aae-4eb1-9509-df8aa672175d", 00:29:56.508 "is_configured": true, 00:29:56.508 "data_offset": 0, 00:29:56.508 "data_size": 65536 00:29:56.508 }, 00:29:56.508 { 00:29:56.508 "name": "BaseBdev4", 00:29:56.508 "uuid": "4f510a7a-c2f6-4f78-99d3-d27ef3a42e7b", 00:29:56.508 "is_configured": true, 00:29:56.508 "data_offset": 0, 00:29:56.508 "data_size": 65536 00:29:56.508 } 00:29:56.508 ] 00:29:56.508 }' 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:56.508 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:56.767 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:56.767 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:29:56.767 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:56.767 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:56.767 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:56.767 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:29:56.767 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:56.767 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:56.767 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:56.767 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:29:56.767 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:56.767 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u bcfd4a30-5d71-4e62-8e02-cc2bebbed34b 00:29:56.767 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:56.767 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:56.767 [2024-11-20 13:50:03.667804] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:29:56.767 [2024-11-20 13:50:03.667841] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:29:56.767 [2024-11-20 13:50:03.667849] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:29:56.768 [2024-11-20 13:50:03.668103] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:29:56.768 [2024-11-20 13:50:03.668234] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:29:56.768 [2024-11-20 13:50:03.668247] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:29:56.768 [2024-11-20 13:50:03.668463] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:56.768 NewBaseBdev 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:56.768 [ 00:29:56.768 { 00:29:56.768 "name": "NewBaseBdev", 00:29:56.768 "aliases": [ 00:29:56.768 "bcfd4a30-5d71-4e62-8e02-cc2bebbed34b" 00:29:56.768 ], 00:29:56.768 "product_name": "Malloc disk", 00:29:56.768 "block_size": 512, 00:29:56.768 "num_blocks": 65536, 00:29:56.768 "uuid": "bcfd4a30-5d71-4e62-8e02-cc2bebbed34b", 00:29:56.768 "assigned_rate_limits": { 00:29:56.768 "rw_ios_per_sec": 0, 00:29:56.768 "rw_mbytes_per_sec": 0, 00:29:56.768 "r_mbytes_per_sec": 0, 00:29:56.768 "w_mbytes_per_sec": 0 00:29:56.768 }, 00:29:56.768 "claimed": true, 00:29:56.768 "claim_type": "exclusive_write", 00:29:56.768 "zoned": false, 00:29:56.768 "supported_io_types": { 00:29:56.768 "read": true, 00:29:56.768 "write": true, 00:29:56.768 "unmap": true, 00:29:56.768 "flush": true, 00:29:56.768 "reset": true, 00:29:56.768 "nvme_admin": false, 00:29:56.768 "nvme_io": false, 00:29:56.768 "nvme_io_md": false, 00:29:56.768 "write_zeroes": true, 00:29:56.768 "zcopy": true, 00:29:56.768 "get_zone_info": false, 00:29:56.768 "zone_management": false, 00:29:56.768 "zone_append": false, 00:29:56.768 "compare": false, 00:29:56.768 "compare_and_write": false, 00:29:56.768 "abort": true, 00:29:56.768 "seek_hole": false, 00:29:56.768 "seek_data": false, 00:29:56.768 "copy": true, 00:29:56.768 "nvme_iov_md": false 00:29:56.768 }, 00:29:56.768 "memory_domains": [ 00:29:56.768 { 00:29:56.768 "dma_device_id": "system", 00:29:56.768 "dma_device_type": 1 00:29:56.768 }, 00:29:56.768 { 00:29:56.768 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:56.768 "dma_device_type": 2 00:29:56.768 } 00:29:56.768 ], 00:29:56.768 "driver_specific": {} 00:29:56.768 } 00:29:56.768 ] 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:56.768 "name": "Existed_Raid", 00:29:56.768 "uuid": "d8a4db3d-f73d-43c0-ae3d-e1afb5db0052", 00:29:56.768 "strip_size_kb": 0, 00:29:56.768 "state": "online", 00:29:56.768 "raid_level": "raid1", 00:29:56.768 "superblock": false, 00:29:56.768 "num_base_bdevs": 4, 00:29:56.768 "num_base_bdevs_discovered": 4, 00:29:56.768 "num_base_bdevs_operational": 4, 00:29:56.768 "base_bdevs_list": [ 00:29:56.768 { 00:29:56.768 "name": "NewBaseBdev", 00:29:56.768 "uuid": "bcfd4a30-5d71-4e62-8e02-cc2bebbed34b", 00:29:56.768 "is_configured": true, 00:29:56.768 "data_offset": 0, 00:29:56.768 "data_size": 65536 00:29:56.768 }, 00:29:56.768 { 00:29:56.768 "name": "BaseBdev2", 00:29:56.768 "uuid": "45e48512-b083-460d-bd93-f1cae8ce40cf", 00:29:56.768 "is_configured": true, 00:29:56.768 "data_offset": 0, 00:29:56.768 "data_size": 65536 00:29:56.768 }, 00:29:56.768 { 00:29:56.768 "name": "BaseBdev3", 00:29:56.768 "uuid": "564616e3-0aae-4eb1-9509-df8aa672175d", 00:29:56.768 "is_configured": true, 00:29:56.768 "data_offset": 0, 00:29:56.768 "data_size": 65536 00:29:56.768 }, 00:29:56.768 { 00:29:56.768 "name": "BaseBdev4", 00:29:56.768 "uuid": "4f510a7a-c2f6-4f78-99d3-d27ef3a42e7b", 00:29:56.768 "is_configured": true, 00:29:56.768 "data_offset": 0, 00:29:56.768 "data_size": 65536 00:29:56.768 } 00:29:56.768 ] 00:29:56.768 }' 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:56.768 13:50:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:57.027 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:29:57.027 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:29:57.027 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:29:57.027 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:29:57.027 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:29:57.027 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:29:57.027 13:50:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:29:57.027 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:57.027 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:29:57.027 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:57.027 [2024-11-20 13:50:04.008278] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:57.027 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:57.027 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:57.027 "name": "Existed_Raid", 00:29:57.027 "aliases": [ 00:29:57.027 "d8a4db3d-f73d-43c0-ae3d-e1afb5db0052" 00:29:57.027 ], 00:29:57.027 "product_name": "Raid Volume", 00:29:57.027 "block_size": 512, 00:29:57.027 "num_blocks": 65536, 00:29:57.027 "uuid": "d8a4db3d-f73d-43c0-ae3d-e1afb5db0052", 00:29:57.027 "assigned_rate_limits": { 00:29:57.027 "rw_ios_per_sec": 0, 00:29:57.027 "rw_mbytes_per_sec": 0, 00:29:57.027 "r_mbytes_per_sec": 0, 00:29:57.027 "w_mbytes_per_sec": 0 00:29:57.027 }, 00:29:57.027 "claimed": false, 00:29:57.027 "zoned": false, 00:29:57.027 "supported_io_types": { 00:29:57.027 "read": true, 00:29:57.027 "write": true, 00:29:57.027 "unmap": false, 00:29:57.027 "flush": false, 00:29:57.027 "reset": true, 00:29:57.027 "nvme_admin": false, 00:29:57.027 "nvme_io": false, 00:29:57.027 "nvme_io_md": false, 00:29:57.027 "write_zeroes": true, 00:29:57.027 "zcopy": false, 00:29:57.027 "get_zone_info": false, 00:29:57.027 "zone_management": false, 00:29:57.027 "zone_append": false, 00:29:57.027 "compare": false, 00:29:57.027 "compare_and_write": false, 00:29:57.027 "abort": false, 00:29:57.027 "seek_hole": false, 00:29:57.027 "seek_data": false, 00:29:57.027 "copy": false, 00:29:57.027 "nvme_iov_md": false 00:29:57.027 }, 00:29:57.027 "memory_domains": [ 00:29:57.027 { 00:29:57.027 "dma_device_id": "system", 00:29:57.027 "dma_device_type": 1 00:29:57.027 }, 00:29:57.027 { 00:29:57.027 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:57.027 "dma_device_type": 2 00:29:57.027 }, 00:29:57.027 { 00:29:57.027 "dma_device_id": "system", 00:29:57.027 "dma_device_type": 1 00:29:57.027 }, 00:29:57.027 { 00:29:57.027 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:57.027 "dma_device_type": 2 00:29:57.027 }, 00:29:57.027 { 00:29:57.027 "dma_device_id": "system", 00:29:57.027 "dma_device_type": 1 00:29:57.027 }, 00:29:57.027 { 00:29:57.027 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:57.027 "dma_device_type": 2 00:29:57.027 }, 00:29:57.027 { 00:29:57.027 "dma_device_id": "system", 00:29:57.027 "dma_device_type": 1 00:29:57.027 }, 00:29:57.027 { 00:29:57.027 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:57.027 "dma_device_type": 2 00:29:57.027 } 00:29:57.027 ], 00:29:57.027 "driver_specific": { 00:29:57.027 "raid": { 00:29:57.027 "uuid": "d8a4db3d-f73d-43c0-ae3d-e1afb5db0052", 00:29:57.027 "strip_size_kb": 0, 00:29:57.028 "state": "online", 00:29:57.028 "raid_level": "raid1", 00:29:57.028 "superblock": false, 00:29:57.028 "num_base_bdevs": 4, 00:29:57.028 "num_base_bdevs_discovered": 4, 00:29:57.028 "num_base_bdevs_operational": 4, 00:29:57.028 "base_bdevs_list": [ 00:29:57.028 { 00:29:57.028 "name": "NewBaseBdev", 00:29:57.028 "uuid": "bcfd4a30-5d71-4e62-8e02-cc2bebbed34b", 00:29:57.028 "is_configured": true, 00:29:57.028 "data_offset": 0, 00:29:57.028 "data_size": 65536 00:29:57.028 }, 00:29:57.028 { 00:29:57.028 "name": "BaseBdev2", 00:29:57.028 "uuid": "45e48512-b083-460d-bd93-f1cae8ce40cf", 00:29:57.028 "is_configured": true, 00:29:57.028 "data_offset": 0, 00:29:57.028 "data_size": 65536 00:29:57.028 }, 00:29:57.028 { 00:29:57.028 "name": "BaseBdev3", 00:29:57.028 "uuid": "564616e3-0aae-4eb1-9509-df8aa672175d", 00:29:57.028 "is_configured": true, 00:29:57.028 "data_offset": 0, 00:29:57.028 "data_size": 65536 00:29:57.028 }, 00:29:57.028 { 00:29:57.028 "name": "BaseBdev4", 00:29:57.028 "uuid": "4f510a7a-c2f6-4f78-99d3-d27ef3a42e7b", 00:29:57.028 "is_configured": true, 00:29:57.028 "data_offset": 0, 00:29:57.028 "data_size": 65536 00:29:57.028 } 00:29:57.028 ] 00:29:57.028 } 00:29:57.028 } 00:29:57.028 }' 00:29:57.028 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:29:57.028 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:29:57.028 BaseBdev2 00:29:57.028 BaseBdev3 00:29:57.028 BaseBdev4' 00:29:57.028 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:57.285 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:29:57.285 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:57.285 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:29:57.285 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:57.285 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:57.285 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:57.285 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:57.285 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:57.285 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:57.285 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:57.285 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:29:57.285 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:57.285 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:57.285 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:57.285 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:57.285 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:57.285 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:57.286 [2024-11-20 13:50:04.235959] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:29:57.286 [2024-11-20 13:50:04.235999] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:29:57.286 [2024-11-20 13:50:04.236084] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:57.286 [2024-11-20 13:50:04.236354] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:29:57.286 [2024-11-20 13:50:04.236366] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 71196 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 71196 ']' 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 71196 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71196 00:29:57.286 killing process with pid 71196 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71196' 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 71196 00:29:57.286 [2024-11-20 13:50:04.265793] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:29:57.286 13:50:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 71196 00:29:57.543 [2024-11-20 13:50:04.471799] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:29:58.108 13:50:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:29:58.109 00:29:58.109 real 0m8.364s 00:29:58.109 user 0m13.445s 00:29:58.109 sys 0m1.438s 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:58.109 ************************************ 00:29:58.109 END TEST raid_state_function_test 00:29:58.109 ************************************ 00:29:58.109 13:50:05 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 4 true 00:29:58.109 13:50:05 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:29:58.109 13:50:05 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:29:58.109 13:50:05 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:29:58.109 ************************************ 00:29:58.109 START TEST raid_state_function_test_sb 00:29:58.109 ************************************ 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 4 true 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:58.109 Process raid pid: 71834 00:29:58.109 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=71834 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 71834' 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 71834 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 71834 ']' 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:29:58.109 13:50:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:58.367 [2024-11-20 13:50:05.223654] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:29:58.367 [2024-11-20 13:50:05.223806] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:29:58.367 [2024-11-20 13:50:05.391703] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:58.625 [2024-11-20 13:50:05.498176] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:29:58.625 [2024-11-20 13:50:05.627762] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:58.625 [2024-11-20 13:50:05.627810] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:59.191 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:29:59.191 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:29:59.191 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:29:59.191 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:59.191 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:59.191 [2024-11-20 13:50:06.055860] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:29:59.191 [2024-11-20 13:50:06.055924] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:29:59.191 [2024-11-20 13:50:06.055934] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:29:59.191 [2024-11-20 13:50:06.055942] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:29:59.191 [2024-11-20 13:50:06.055948] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:29:59.191 [2024-11-20 13:50:06.055955] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:29:59.191 [2024-11-20 13:50:06.055961] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:29:59.191 [2024-11-20 13:50:06.055968] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:29:59.191 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:59.191 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:29:59.191 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:59.191 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:59.191 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:29:59.191 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:29:59.191 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:59.191 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:59.191 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:59.191 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:59.191 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:59.191 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:59.191 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:59.191 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:59.191 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:59.191 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:59.191 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:59.191 "name": "Existed_Raid", 00:29:59.191 "uuid": "0795aec7-98c5-4a25-a2bd-b940400ab90c", 00:29:59.191 "strip_size_kb": 0, 00:29:59.191 "state": "configuring", 00:29:59.191 "raid_level": "raid1", 00:29:59.191 "superblock": true, 00:29:59.191 "num_base_bdevs": 4, 00:29:59.191 "num_base_bdevs_discovered": 0, 00:29:59.191 "num_base_bdevs_operational": 4, 00:29:59.191 "base_bdevs_list": [ 00:29:59.191 { 00:29:59.191 "name": "BaseBdev1", 00:29:59.191 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:59.191 "is_configured": false, 00:29:59.191 "data_offset": 0, 00:29:59.191 "data_size": 0 00:29:59.191 }, 00:29:59.191 { 00:29:59.191 "name": "BaseBdev2", 00:29:59.191 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:59.191 "is_configured": false, 00:29:59.191 "data_offset": 0, 00:29:59.191 "data_size": 0 00:29:59.191 }, 00:29:59.191 { 00:29:59.191 "name": "BaseBdev3", 00:29:59.191 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:59.191 "is_configured": false, 00:29:59.191 "data_offset": 0, 00:29:59.191 "data_size": 0 00:29:59.191 }, 00:29:59.191 { 00:29:59.191 "name": "BaseBdev4", 00:29:59.191 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:59.191 "is_configured": false, 00:29:59.191 "data_offset": 0, 00:29:59.191 "data_size": 0 00:29:59.191 } 00:29:59.191 ] 00:29:59.191 }' 00:29:59.191 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:59.191 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:59.450 [2024-11-20 13:50:06.363853] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:29:59.450 [2024-11-20 13:50:06.363897] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:59.450 [2024-11-20 13:50:06.371847] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:29:59.450 [2024-11-20 13:50:06.371887] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:29:59.450 [2024-11-20 13:50:06.371895] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:29:59.450 [2024-11-20 13:50:06.371902] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:29:59.450 [2024-11-20 13:50:06.371907] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:29:59.450 [2024-11-20 13:50:06.371915] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:29:59.450 [2024-11-20 13:50:06.371919] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:29:59.450 [2024-11-20 13:50:06.371927] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:59.450 [2024-11-20 13:50:06.402567] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:29:59.450 BaseBdev1 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:59.450 [ 00:29:59.450 { 00:29:59.450 "name": "BaseBdev1", 00:29:59.450 "aliases": [ 00:29:59.450 "e86ec302-1a3b-435c-a26e-0baf4c0ef34f" 00:29:59.450 ], 00:29:59.450 "product_name": "Malloc disk", 00:29:59.450 "block_size": 512, 00:29:59.450 "num_blocks": 65536, 00:29:59.450 "uuid": "e86ec302-1a3b-435c-a26e-0baf4c0ef34f", 00:29:59.450 "assigned_rate_limits": { 00:29:59.450 "rw_ios_per_sec": 0, 00:29:59.450 "rw_mbytes_per_sec": 0, 00:29:59.450 "r_mbytes_per_sec": 0, 00:29:59.450 "w_mbytes_per_sec": 0 00:29:59.450 }, 00:29:59.450 "claimed": true, 00:29:59.450 "claim_type": "exclusive_write", 00:29:59.450 "zoned": false, 00:29:59.450 "supported_io_types": { 00:29:59.450 "read": true, 00:29:59.450 "write": true, 00:29:59.450 "unmap": true, 00:29:59.450 "flush": true, 00:29:59.450 "reset": true, 00:29:59.450 "nvme_admin": false, 00:29:59.450 "nvme_io": false, 00:29:59.450 "nvme_io_md": false, 00:29:59.450 "write_zeroes": true, 00:29:59.450 "zcopy": true, 00:29:59.450 "get_zone_info": false, 00:29:59.450 "zone_management": false, 00:29:59.450 "zone_append": false, 00:29:59.450 "compare": false, 00:29:59.450 "compare_and_write": false, 00:29:59.450 "abort": true, 00:29:59.450 "seek_hole": false, 00:29:59.450 "seek_data": false, 00:29:59.450 "copy": true, 00:29:59.450 "nvme_iov_md": false 00:29:59.450 }, 00:29:59.450 "memory_domains": [ 00:29:59.450 { 00:29:59.450 "dma_device_id": "system", 00:29:59.450 "dma_device_type": 1 00:29:59.450 }, 00:29:59.450 { 00:29:59.450 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:59.450 "dma_device_type": 2 00:29:59.450 } 00:29:59.450 ], 00:29:59.450 "driver_specific": {} 00:29:59.450 } 00:29:59.450 ] 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:59.450 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:59.451 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:59.451 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:59.451 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:59.451 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:59.451 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:59.451 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:59.451 "name": "Existed_Raid", 00:29:59.451 "uuid": "f618c154-eeb5-4ef0-9d41-a4e78e07853c", 00:29:59.451 "strip_size_kb": 0, 00:29:59.451 "state": "configuring", 00:29:59.451 "raid_level": "raid1", 00:29:59.451 "superblock": true, 00:29:59.451 "num_base_bdevs": 4, 00:29:59.451 "num_base_bdevs_discovered": 1, 00:29:59.451 "num_base_bdevs_operational": 4, 00:29:59.451 "base_bdevs_list": [ 00:29:59.451 { 00:29:59.451 "name": "BaseBdev1", 00:29:59.451 "uuid": "e86ec302-1a3b-435c-a26e-0baf4c0ef34f", 00:29:59.451 "is_configured": true, 00:29:59.451 "data_offset": 2048, 00:29:59.451 "data_size": 63488 00:29:59.451 }, 00:29:59.451 { 00:29:59.451 "name": "BaseBdev2", 00:29:59.451 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:59.451 "is_configured": false, 00:29:59.451 "data_offset": 0, 00:29:59.451 "data_size": 0 00:29:59.451 }, 00:29:59.451 { 00:29:59.451 "name": "BaseBdev3", 00:29:59.451 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:59.451 "is_configured": false, 00:29:59.451 "data_offset": 0, 00:29:59.451 "data_size": 0 00:29:59.451 }, 00:29:59.451 { 00:29:59.451 "name": "BaseBdev4", 00:29:59.451 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:59.451 "is_configured": false, 00:29:59.451 "data_offset": 0, 00:29:59.451 "data_size": 0 00:29:59.451 } 00:29:59.451 ] 00:29:59.451 }' 00:29:59.451 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:59.451 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:59.709 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:29:59.709 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:59.709 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:59.709 [2024-11-20 13:50:06.750678] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:29:59.709 [2024-11-20 13:50:06.750867] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:29:59.709 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:59.709 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:29:59.709 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:59.709 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:59.709 [2024-11-20 13:50:06.758723] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:29:59.709 [2024-11-20 13:50:06.760417] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:29:59.709 [2024-11-20 13:50:06.760467] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:29:59.709 [2024-11-20 13:50:06.760475] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:29:59.709 [2024-11-20 13:50:06.760485] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:29:59.709 [2024-11-20 13:50:06.760491] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:29:59.709 [2024-11-20 13:50:06.760498] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:29:59.709 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:59.709 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:29:59.709 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:29:59.709 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:29:59.709 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:59.709 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:59.709 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:29:59.709 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:29:59.709 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:29:59.709 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:59.709 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:59.709 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:59.709 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:59.967 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:59.967 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:59.967 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:59.967 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:59.967 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:59.967 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:59.967 "name": "Existed_Raid", 00:29:59.967 "uuid": "f2e67983-6902-448a-ade7-27c2797b46c6", 00:29:59.967 "strip_size_kb": 0, 00:29:59.967 "state": "configuring", 00:29:59.967 "raid_level": "raid1", 00:29:59.967 "superblock": true, 00:29:59.967 "num_base_bdevs": 4, 00:29:59.967 "num_base_bdevs_discovered": 1, 00:29:59.967 "num_base_bdevs_operational": 4, 00:29:59.967 "base_bdevs_list": [ 00:29:59.967 { 00:29:59.967 "name": "BaseBdev1", 00:29:59.967 "uuid": "e86ec302-1a3b-435c-a26e-0baf4c0ef34f", 00:29:59.967 "is_configured": true, 00:29:59.967 "data_offset": 2048, 00:29:59.967 "data_size": 63488 00:29:59.967 }, 00:29:59.967 { 00:29:59.967 "name": "BaseBdev2", 00:29:59.967 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:59.967 "is_configured": false, 00:29:59.967 "data_offset": 0, 00:29:59.967 "data_size": 0 00:29:59.967 }, 00:29:59.967 { 00:29:59.967 "name": "BaseBdev3", 00:29:59.967 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:59.967 "is_configured": false, 00:29:59.967 "data_offset": 0, 00:29:59.967 "data_size": 0 00:29:59.967 }, 00:29:59.967 { 00:29:59.967 "name": "BaseBdev4", 00:29:59.967 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:59.967 "is_configured": false, 00:29:59.967 "data_offset": 0, 00:29:59.967 "data_size": 0 00:29:59.967 } 00:29:59.967 ] 00:29:59.967 }' 00:29:59.967 13:50:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:59.967 13:50:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:00.225 [2024-11-20 13:50:07.087540] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:30:00.225 BaseBdev2 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:00.225 [ 00:30:00.225 { 00:30:00.225 "name": "BaseBdev2", 00:30:00.225 "aliases": [ 00:30:00.225 "94cc7e0c-6b88-4a06-86fd-eb13705fb48c" 00:30:00.225 ], 00:30:00.225 "product_name": "Malloc disk", 00:30:00.225 "block_size": 512, 00:30:00.225 "num_blocks": 65536, 00:30:00.225 "uuid": "94cc7e0c-6b88-4a06-86fd-eb13705fb48c", 00:30:00.225 "assigned_rate_limits": { 00:30:00.225 "rw_ios_per_sec": 0, 00:30:00.225 "rw_mbytes_per_sec": 0, 00:30:00.225 "r_mbytes_per_sec": 0, 00:30:00.225 "w_mbytes_per_sec": 0 00:30:00.225 }, 00:30:00.225 "claimed": true, 00:30:00.225 "claim_type": "exclusive_write", 00:30:00.225 "zoned": false, 00:30:00.225 "supported_io_types": { 00:30:00.225 "read": true, 00:30:00.225 "write": true, 00:30:00.225 "unmap": true, 00:30:00.225 "flush": true, 00:30:00.225 "reset": true, 00:30:00.225 "nvme_admin": false, 00:30:00.225 "nvme_io": false, 00:30:00.225 "nvme_io_md": false, 00:30:00.225 "write_zeroes": true, 00:30:00.225 "zcopy": true, 00:30:00.225 "get_zone_info": false, 00:30:00.225 "zone_management": false, 00:30:00.225 "zone_append": false, 00:30:00.225 "compare": false, 00:30:00.225 "compare_and_write": false, 00:30:00.225 "abort": true, 00:30:00.225 "seek_hole": false, 00:30:00.225 "seek_data": false, 00:30:00.225 "copy": true, 00:30:00.225 "nvme_iov_md": false 00:30:00.225 }, 00:30:00.225 "memory_domains": [ 00:30:00.225 { 00:30:00.225 "dma_device_id": "system", 00:30:00.225 "dma_device_type": 1 00:30:00.225 }, 00:30:00.225 { 00:30:00.225 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:00.225 "dma_device_type": 2 00:30:00.225 } 00:30:00.225 ], 00:30:00.225 "driver_specific": {} 00:30:00.225 } 00:30:00.225 ] 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:00.225 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:00.225 "name": "Existed_Raid", 00:30:00.225 "uuid": "f2e67983-6902-448a-ade7-27c2797b46c6", 00:30:00.225 "strip_size_kb": 0, 00:30:00.225 "state": "configuring", 00:30:00.225 "raid_level": "raid1", 00:30:00.225 "superblock": true, 00:30:00.225 "num_base_bdevs": 4, 00:30:00.225 "num_base_bdevs_discovered": 2, 00:30:00.225 "num_base_bdevs_operational": 4, 00:30:00.225 "base_bdevs_list": [ 00:30:00.225 { 00:30:00.225 "name": "BaseBdev1", 00:30:00.225 "uuid": "e86ec302-1a3b-435c-a26e-0baf4c0ef34f", 00:30:00.225 "is_configured": true, 00:30:00.225 "data_offset": 2048, 00:30:00.225 "data_size": 63488 00:30:00.225 }, 00:30:00.225 { 00:30:00.225 "name": "BaseBdev2", 00:30:00.225 "uuid": "94cc7e0c-6b88-4a06-86fd-eb13705fb48c", 00:30:00.225 "is_configured": true, 00:30:00.225 "data_offset": 2048, 00:30:00.225 "data_size": 63488 00:30:00.225 }, 00:30:00.226 { 00:30:00.226 "name": "BaseBdev3", 00:30:00.226 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:00.226 "is_configured": false, 00:30:00.226 "data_offset": 0, 00:30:00.226 "data_size": 0 00:30:00.226 }, 00:30:00.226 { 00:30:00.226 "name": "BaseBdev4", 00:30:00.226 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:00.226 "is_configured": false, 00:30:00.226 "data_offset": 0, 00:30:00.226 "data_size": 0 00:30:00.226 } 00:30:00.226 ] 00:30:00.226 }' 00:30:00.226 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:00.226 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:00.483 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:30:00.483 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:00.483 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:00.483 [2024-11-20 13:50:07.460322] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:30:00.483 BaseBdev3 00:30:00.483 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:00.484 [ 00:30:00.484 { 00:30:00.484 "name": "BaseBdev3", 00:30:00.484 "aliases": [ 00:30:00.484 "c12dfdd2-85fb-4e25-a522-123b70b5eb41" 00:30:00.484 ], 00:30:00.484 "product_name": "Malloc disk", 00:30:00.484 "block_size": 512, 00:30:00.484 "num_blocks": 65536, 00:30:00.484 "uuid": "c12dfdd2-85fb-4e25-a522-123b70b5eb41", 00:30:00.484 "assigned_rate_limits": { 00:30:00.484 "rw_ios_per_sec": 0, 00:30:00.484 "rw_mbytes_per_sec": 0, 00:30:00.484 "r_mbytes_per_sec": 0, 00:30:00.484 "w_mbytes_per_sec": 0 00:30:00.484 }, 00:30:00.484 "claimed": true, 00:30:00.484 "claim_type": "exclusive_write", 00:30:00.484 "zoned": false, 00:30:00.484 "supported_io_types": { 00:30:00.484 "read": true, 00:30:00.484 "write": true, 00:30:00.484 "unmap": true, 00:30:00.484 "flush": true, 00:30:00.484 "reset": true, 00:30:00.484 "nvme_admin": false, 00:30:00.484 "nvme_io": false, 00:30:00.484 "nvme_io_md": false, 00:30:00.484 "write_zeroes": true, 00:30:00.484 "zcopy": true, 00:30:00.484 "get_zone_info": false, 00:30:00.484 "zone_management": false, 00:30:00.484 "zone_append": false, 00:30:00.484 "compare": false, 00:30:00.484 "compare_and_write": false, 00:30:00.484 "abort": true, 00:30:00.484 "seek_hole": false, 00:30:00.484 "seek_data": false, 00:30:00.484 "copy": true, 00:30:00.484 "nvme_iov_md": false 00:30:00.484 }, 00:30:00.484 "memory_domains": [ 00:30:00.484 { 00:30:00.484 "dma_device_id": "system", 00:30:00.484 "dma_device_type": 1 00:30:00.484 }, 00:30:00.484 { 00:30:00.484 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:00.484 "dma_device_type": 2 00:30:00.484 } 00:30:00.484 ], 00:30:00.484 "driver_specific": {} 00:30:00.484 } 00:30:00.484 ] 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:00.484 "name": "Existed_Raid", 00:30:00.484 "uuid": "f2e67983-6902-448a-ade7-27c2797b46c6", 00:30:00.484 "strip_size_kb": 0, 00:30:00.484 "state": "configuring", 00:30:00.484 "raid_level": "raid1", 00:30:00.484 "superblock": true, 00:30:00.484 "num_base_bdevs": 4, 00:30:00.484 "num_base_bdevs_discovered": 3, 00:30:00.484 "num_base_bdevs_operational": 4, 00:30:00.484 "base_bdevs_list": [ 00:30:00.484 { 00:30:00.484 "name": "BaseBdev1", 00:30:00.484 "uuid": "e86ec302-1a3b-435c-a26e-0baf4c0ef34f", 00:30:00.484 "is_configured": true, 00:30:00.484 "data_offset": 2048, 00:30:00.484 "data_size": 63488 00:30:00.484 }, 00:30:00.484 { 00:30:00.484 "name": "BaseBdev2", 00:30:00.484 "uuid": "94cc7e0c-6b88-4a06-86fd-eb13705fb48c", 00:30:00.484 "is_configured": true, 00:30:00.484 "data_offset": 2048, 00:30:00.484 "data_size": 63488 00:30:00.484 }, 00:30:00.484 { 00:30:00.484 "name": "BaseBdev3", 00:30:00.484 "uuid": "c12dfdd2-85fb-4e25-a522-123b70b5eb41", 00:30:00.484 "is_configured": true, 00:30:00.484 "data_offset": 2048, 00:30:00.484 "data_size": 63488 00:30:00.484 }, 00:30:00.484 { 00:30:00.484 "name": "BaseBdev4", 00:30:00.484 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:00.484 "is_configured": false, 00:30:00.484 "data_offset": 0, 00:30:00.484 "data_size": 0 00:30:00.484 } 00:30:00.484 ] 00:30:00.484 }' 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:00.484 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:00.746 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:30:00.746 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:00.746 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:01.003 [2024-11-20 13:50:07.817345] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:30:01.003 [2024-11-20 13:50:07.817739] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:30:01.003 [2024-11-20 13:50:07.817825] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:30:01.003 BaseBdev4 00:30:01.003 [2024-11-20 13:50:07.818116] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:30:01.003 [2024-11-20 13:50:07.818255] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:30:01.003 [2024-11-20 13:50:07.818266] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:30:01.003 [2024-11-20 13:50:07.818391] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:01.003 [ 00:30:01.003 { 00:30:01.003 "name": "BaseBdev4", 00:30:01.003 "aliases": [ 00:30:01.003 "e3b54894-4037-4578-b356-53edb639a84c" 00:30:01.003 ], 00:30:01.003 "product_name": "Malloc disk", 00:30:01.003 "block_size": 512, 00:30:01.003 "num_blocks": 65536, 00:30:01.003 "uuid": "e3b54894-4037-4578-b356-53edb639a84c", 00:30:01.003 "assigned_rate_limits": { 00:30:01.003 "rw_ios_per_sec": 0, 00:30:01.003 "rw_mbytes_per_sec": 0, 00:30:01.003 "r_mbytes_per_sec": 0, 00:30:01.003 "w_mbytes_per_sec": 0 00:30:01.003 }, 00:30:01.003 "claimed": true, 00:30:01.003 "claim_type": "exclusive_write", 00:30:01.003 "zoned": false, 00:30:01.003 "supported_io_types": { 00:30:01.003 "read": true, 00:30:01.003 "write": true, 00:30:01.003 "unmap": true, 00:30:01.003 "flush": true, 00:30:01.003 "reset": true, 00:30:01.003 "nvme_admin": false, 00:30:01.003 "nvme_io": false, 00:30:01.003 "nvme_io_md": false, 00:30:01.003 "write_zeroes": true, 00:30:01.003 "zcopy": true, 00:30:01.003 "get_zone_info": false, 00:30:01.003 "zone_management": false, 00:30:01.003 "zone_append": false, 00:30:01.003 "compare": false, 00:30:01.003 "compare_and_write": false, 00:30:01.003 "abort": true, 00:30:01.003 "seek_hole": false, 00:30:01.003 "seek_data": false, 00:30:01.003 "copy": true, 00:30:01.003 "nvme_iov_md": false 00:30:01.003 }, 00:30:01.003 "memory_domains": [ 00:30:01.003 { 00:30:01.003 "dma_device_id": "system", 00:30:01.003 "dma_device_type": 1 00:30:01.003 }, 00:30:01.003 { 00:30:01.003 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:01.003 "dma_device_type": 2 00:30:01.003 } 00:30:01.003 ], 00:30:01.003 "driver_specific": {} 00:30:01.003 } 00:30:01.003 ] 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:01.003 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:01.003 "name": "Existed_Raid", 00:30:01.003 "uuid": "f2e67983-6902-448a-ade7-27c2797b46c6", 00:30:01.003 "strip_size_kb": 0, 00:30:01.003 "state": "online", 00:30:01.003 "raid_level": "raid1", 00:30:01.003 "superblock": true, 00:30:01.003 "num_base_bdevs": 4, 00:30:01.003 "num_base_bdevs_discovered": 4, 00:30:01.003 "num_base_bdevs_operational": 4, 00:30:01.003 "base_bdevs_list": [ 00:30:01.003 { 00:30:01.003 "name": "BaseBdev1", 00:30:01.003 "uuid": "e86ec302-1a3b-435c-a26e-0baf4c0ef34f", 00:30:01.003 "is_configured": true, 00:30:01.003 "data_offset": 2048, 00:30:01.003 "data_size": 63488 00:30:01.003 }, 00:30:01.003 { 00:30:01.003 "name": "BaseBdev2", 00:30:01.003 "uuid": "94cc7e0c-6b88-4a06-86fd-eb13705fb48c", 00:30:01.003 "is_configured": true, 00:30:01.003 "data_offset": 2048, 00:30:01.003 "data_size": 63488 00:30:01.003 }, 00:30:01.003 { 00:30:01.003 "name": "BaseBdev3", 00:30:01.003 "uuid": "c12dfdd2-85fb-4e25-a522-123b70b5eb41", 00:30:01.003 "is_configured": true, 00:30:01.003 "data_offset": 2048, 00:30:01.003 "data_size": 63488 00:30:01.003 }, 00:30:01.003 { 00:30:01.003 "name": "BaseBdev4", 00:30:01.004 "uuid": "e3b54894-4037-4578-b356-53edb639a84c", 00:30:01.004 "is_configured": true, 00:30:01.004 "data_offset": 2048, 00:30:01.004 "data_size": 63488 00:30:01.004 } 00:30:01.004 ] 00:30:01.004 }' 00:30:01.004 13:50:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:01.004 13:50:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:01.261 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:30:01.261 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:30:01.261 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:30:01.261 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:30:01.261 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:30:01.261 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:30:01.261 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:30:01.261 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:30:01.261 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:01.261 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:01.261 [2024-11-20 13:50:08.165799] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:30:01.261 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:01.261 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:30:01.261 "name": "Existed_Raid", 00:30:01.261 "aliases": [ 00:30:01.261 "f2e67983-6902-448a-ade7-27c2797b46c6" 00:30:01.261 ], 00:30:01.261 "product_name": "Raid Volume", 00:30:01.261 "block_size": 512, 00:30:01.261 "num_blocks": 63488, 00:30:01.261 "uuid": "f2e67983-6902-448a-ade7-27c2797b46c6", 00:30:01.261 "assigned_rate_limits": { 00:30:01.261 "rw_ios_per_sec": 0, 00:30:01.261 "rw_mbytes_per_sec": 0, 00:30:01.261 "r_mbytes_per_sec": 0, 00:30:01.261 "w_mbytes_per_sec": 0 00:30:01.261 }, 00:30:01.261 "claimed": false, 00:30:01.261 "zoned": false, 00:30:01.261 "supported_io_types": { 00:30:01.261 "read": true, 00:30:01.261 "write": true, 00:30:01.261 "unmap": false, 00:30:01.261 "flush": false, 00:30:01.261 "reset": true, 00:30:01.261 "nvme_admin": false, 00:30:01.261 "nvme_io": false, 00:30:01.261 "nvme_io_md": false, 00:30:01.261 "write_zeroes": true, 00:30:01.262 "zcopy": false, 00:30:01.262 "get_zone_info": false, 00:30:01.262 "zone_management": false, 00:30:01.262 "zone_append": false, 00:30:01.262 "compare": false, 00:30:01.262 "compare_and_write": false, 00:30:01.262 "abort": false, 00:30:01.262 "seek_hole": false, 00:30:01.262 "seek_data": false, 00:30:01.262 "copy": false, 00:30:01.262 "nvme_iov_md": false 00:30:01.262 }, 00:30:01.262 "memory_domains": [ 00:30:01.262 { 00:30:01.262 "dma_device_id": "system", 00:30:01.262 "dma_device_type": 1 00:30:01.262 }, 00:30:01.262 { 00:30:01.262 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:01.262 "dma_device_type": 2 00:30:01.262 }, 00:30:01.262 { 00:30:01.262 "dma_device_id": "system", 00:30:01.262 "dma_device_type": 1 00:30:01.262 }, 00:30:01.262 { 00:30:01.262 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:01.262 "dma_device_type": 2 00:30:01.262 }, 00:30:01.262 { 00:30:01.262 "dma_device_id": "system", 00:30:01.262 "dma_device_type": 1 00:30:01.262 }, 00:30:01.262 { 00:30:01.262 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:01.262 "dma_device_type": 2 00:30:01.262 }, 00:30:01.262 { 00:30:01.262 "dma_device_id": "system", 00:30:01.262 "dma_device_type": 1 00:30:01.262 }, 00:30:01.262 { 00:30:01.262 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:01.262 "dma_device_type": 2 00:30:01.262 } 00:30:01.262 ], 00:30:01.262 "driver_specific": { 00:30:01.262 "raid": { 00:30:01.262 "uuid": "f2e67983-6902-448a-ade7-27c2797b46c6", 00:30:01.262 "strip_size_kb": 0, 00:30:01.262 "state": "online", 00:30:01.262 "raid_level": "raid1", 00:30:01.262 "superblock": true, 00:30:01.262 "num_base_bdevs": 4, 00:30:01.262 "num_base_bdevs_discovered": 4, 00:30:01.262 "num_base_bdevs_operational": 4, 00:30:01.262 "base_bdevs_list": [ 00:30:01.262 { 00:30:01.262 "name": "BaseBdev1", 00:30:01.262 "uuid": "e86ec302-1a3b-435c-a26e-0baf4c0ef34f", 00:30:01.262 "is_configured": true, 00:30:01.262 "data_offset": 2048, 00:30:01.262 "data_size": 63488 00:30:01.262 }, 00:30:01.262 { 00:30:01.262 "name": "BaseBdev2", 00:30:01.262 "uuid": "94cc7e0c-6b88-4a06-86fd-eb13705fb48c", 00:30:01.262 "is_configured": true, 00:30:01.262 "data_offset": 2048, 00:30:01.262 "data_size": 63488 00:30:01.262 }, 00:30:01.262 { 00:30:01.262 "name": "BaseBdev3", 00:30:01.262 "uuid": "c12dfdd2-85fb-4e25-a522-123b70b5eb41", 00:30:01.262 "is_configured": true, 00:30:01.262 "data_offset": 2048, 00:30:01.262 "data_size": 63488 00:30:01.262 }, 00:30:01.262 { 00:30:01.262 "name": "BaseBdev4", 00:30:01.262 "uuid": "e3b54894-4037-4578-b356-53edb639a84c", 00:30:01.262 "is_configured": true, 00:30:01.262 "data_offset": 2048, 00:30:01.262 "data_size": 63488 00:30:01.262 } 00:30:01.262 ] 00:30:01.262 } 00:30:01.262 } 00:30:01.262 }' 00:30:01.262 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:30:01.262 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:30:01.262 BaseBdev2 00:30:01.262 BaseBdev3 00:30:01.262 BaseBdev4' 00:30:01.262 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:01.262 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:30:01.262 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:01.262 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:30:01.262 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:01.262 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:01.262 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:01.262 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:01.262 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:01.262 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:01.262 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:01.262 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:01.262 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:30:01.262 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:01.262 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:01.520 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:01.520 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:01.520 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:01.520 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:01.520 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:01.520 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:30:01.520 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:01.520 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:01.520 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:01.520 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:01.520 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:01.520 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:01.520 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:30:01.520 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:01.520 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:01.520 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:01.520 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:01.520 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:01.520 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:01.520 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:30:01.521 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:01.521 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:01.521 [2024-11-20 13:50:08.421595] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:30:01.521 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:01.521 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:30:01.521 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:30:01.521 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:30:01.521 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:30:01.521 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:30:01.521 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:30:01.521 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:01.521 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:01.521 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:01.521 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:01.521 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:30:01.521 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:01.521 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:01.521 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:01.521 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:01.521 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:01.521 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:01.521 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:01.521 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:01.521 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:01.521 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:01.521 "name": "Existed_Raid", 00:30:01.521 "uuid": "f2e67983-6902-448a-ade7-27c2797b46c6", 00:30:01.521 "strip_size_kb": 0, 00:30:01.521 "state": "online", 00:30:01.521 "raid_level": "raid1", 00:30:01.521 "superblock": true, 00:30:01.521 "num_base_bdevs": 4, 00:30:01.521 "num_base_bdevs_discovered": 3, 00:30:01.521 "num_base_bdevs_operational": 3, 00:30:01.521 "base_bdevs_list": [ 00:30:01.521 { 00:30:01.521 "name": null, 00:30:01.521 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:01.521 "is_configured": false, 00:30:01.521 "data_offset": 0, 00:30:01.521 "data_size": 63488 00:30:01.521 }, 00:30:01.521 { 00:30:01.521 "name": "BaseBdev2", 00:30:01.521 "uuid": "94cc7e0c-6b88-4a06-86fd-eb13705fb48c", 00:30:01.521 "is_configured": true, 00:30:01.521 "data_offset": 2048, 00:30:01.521 "data_size": 63488 00:30:01.521 }, 00:30:01.521 { 00:30:01.521 "name": "BaseBdev3", 00:30:01.521 "uuid": "c12dfdd2-85fb-4e25-a522-123b70b5eb41", 00:30:01.521 "is_configured": true, 00:30:01.521 "data_offset": 2048, 00:30:01.521 "data_size": 63488 00:30:01.521 }, 00:30:01.521 { 00:30:01.521 "name": "BaseBdev4", 00:30:01.521 "uuid": "e3b54894-4037-4578-b356-53edb639a84c", 00:30:01.521 "is_configured": true, 00:30:01.521 "data_offset": 2048, 00:30:01.521 "data_size": 63488 00:30:01.521 } 00:30:01.521 ] 00:30:01.521 }' 00:30:01.521 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:01.521 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:01.779 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:30:01.779 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:30:01.779 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:01.779 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:30:01.779 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:01.779 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:01.779 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:01.779 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:30:01.779 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:30:01.779 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:30:01.779 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:01.779 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:01.779 [2024-11-20 13:50:08.828652] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:30:02.037 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:02.037 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:30:02.037 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:30:02.037 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:30:02.037 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:02.037 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:02.037 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:02.037 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:02.037 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:30:02.037 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:30:02.037 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:30:02.037 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:02.037 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:02.037 [2024-11-20 13:50:08.914699] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:30:02.037 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:02.037 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:30:02.037 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:30:02.037 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:30:02.037 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:02.037 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:02.037 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:02.037 13:50:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:02.037 13:50:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:30:02.037 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:30:02.037 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:30:02.037 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:02.037 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:02.037 [2024-11-20 13:50:09.005494] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:30:02.037 [2024-11-20 13:50:09.005600] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:30:02.037 [2024-11-20 13:50:09.056181] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:30:02.037 [2024-11-20 13:50:09.056235] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:30:02.037 [2024-11-20 13:50:09.056245] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:30:02.037 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:02.037 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:30:02.037 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:30:02.037 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:02.037 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:30:02.037 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:02.037 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:02.037 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:02.037 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:30:02.037 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:30:02.037 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:30:02.037 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:30:02.037 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:30:02.037 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:30:02.037 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:02.037 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:02.296 BaseBdev2 00:30:02.296 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:02.296 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:30:02.296 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:30:02.296 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:30:02.296 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:30:02.296 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:30:02.296 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:30:02.296 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:30:02.296 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:02.296 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:02.296 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:02.296 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:30:02.296 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:02.296 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:02.296 [ 00:30:02.296 { 00:30:02.296 "name": "BaseBdev2", 00:30:02.296 "aliases": [ 00:30:02.296 "5eed79cb-fe93-466b-ad51-e3cc15d577e3" 00:30:02.296 ], 00:30:02.296 "product_name": "Malloc disk", 00:30:02.296 "block_size": 512, 00:30:02.296 "num_blocks": 65536, 00:30:02.296 "uuid": "5eed79cb-fe93-466b-ad51-e3cc15d577e3", 00:30:02.296 "assigned_rate_limits": { 00:30:02.296 "rw_ios_per_sec": 0, 00:30:02.296 "rw_mbytes_per_sec": 0, 00:30:02.296 "r_mbytes_per_sec": 0, 00:30:02.296 "w_mbytes_per_sec": 0 00:30:02.296 }, 00:30:02.296 "claimed": false, 00:30:02.296 "zoned": false, 00:30:02.296 "supported_io_types": { 00:30:02.296 "read": true, 00:30:02.296 "write": true, 00:30:02.296 "unmap": true, 00:30:02.296 "flush": true, 00:30:02.296 "reset": true, 00:30:02.296 "nvme_admin": false, 00:30:02.296 "nvme_io": false, 00:30:02.296 "nvme_io_md": false, 00:30:02.296 "write_zeroes": true, 00:30:02.296 "zcopy": true, 00:30:02.296 "get_zone_info": false, 00:30:02.296 "zone_management": false, 00:30:02.296 "zone_append": false, 00:30:02.296 "compare": false, 00:30:02.296 "compare_and_write": false, 00:30:02.296 "abort": true, 00:30:02.296 "seek_hole": false, 00:30:02.296 "seek_data": false, 00:30:02.296 "copy": true, 00:30:02.296 "nvme_iov_md": false 00:30:02.296 }, 00:30:02.296 "memory_domains": [ 00:30:02.296 { 00:30:02.296 "dma_device_id": "system", 00:30:02.296 "dma_device_type": 1 00:30:02.296 }, 00:30:02.296 { 00:30:02.296 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:02.296 "dma_device_type": 2 00:30:02.296 } 00:30:02.296 ], 00:30:02.296 "driver_specific": {} 00:30:02.296 } 00:30:02.296 ] 00:30:02.296 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:02.296 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:30:02.296 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:30:02.296 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:30:02.296 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:30:02.296 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:02.296 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:02.296 BaseBdev3 00:30:02.296 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:02.296 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:30:02.296 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:30:02.296 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:02.297 [ 00:30:02.297 { 00:30:02.297 "name": "BaseBdev3", 00:30:02.297 "aliases": [ 00:30:02.297 "2636ee53-fea7-4da4-ad0e-32e4f6345fd3" 00:30:02.297 ], 00:30:02.297 "product_name": "Malloc disk", 00:30:02.297 "block_size": 512, 00:30:02.297 "num_blocks": 65536, 00:30:02.297 "uuid": "2636ee53-fea7-4da4-ad0e-32e4f6345fd3", 00:30:02.297 "assigned_rate_limits": { 00:30:02.297 "rw_ios_per_sec": 0, 00:30:02.297 "rw_mbytes_per_sec": 0, 00:30:02.297 "r_mbytes_per_sec": 0, 00:30:02.297 "w_mbytes_per_sec": 0 00:30:02.297 }, 00:30:02.297 "claimed": false, 00:30:02.297 "zoned": false, 00:30:02.297 "supported_io_types": { 00:30:02.297 "read": true, 00:30:02.297 "write": true, 00:30:02.297 "unmap": true, 00:30:02.297 "flush": true, 00:30:02.297 "reset": true, 00:30:02.297 "nvme_admin": false, 00:30:02.297 "nvme_io": false, 00:30:02.297 "nvme_io_md": false, 00:30:02.297 "write_zeroes": true, 00:30:02.297 "zcopy": true, 00:30:02.297 "get_zone_info": false, 00:30:02.297 "zone_management": false, 00:30:02.297 "zone_append": false, 00:30:02.297 "compare": false, 00:30:02.297 "compare_and_write": false, 00:30:02.297 "abort": true, 00:30:02.297 "seek_hole": false, 00:30:02.297 "seek_data": false, 00:30:02.297 "copy": true, 00:30:02.297 "nvme_iov_md": false 00:30:02.297 }, 00:30:02.297 "memory_domains": [ 00:30:02.297 { 00:30:02.297 "dma_device_id": "system", 00:30:02.297 "dma_device_type": 1 00:30:02.297 }, 00:30:02.297 { 00:30:02.297 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:02.297 "dma_device_type": 2 00:30:02.297 } 00:30:02.297 ], 00:30:02.297 "driver_specific": {} 00:30:02.297 } 00:30:02.297 ] 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:02.297 BaseBdev4 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:02.297 [ 00:30:02.297 { 00:30:02.297 "name": "BaseBdev4", 00:30:02.297 "aliases": [ 00:30:02.297 "60a8f92e-12f4-4968-8980-adbb931624c6" 00:30:02.297 ], 00:30:02.297 "product_name": "Malloc disk", 00:30:02.297 "block_size": 512, 00:30:02.297 "num_blocks": 65536, 00:30:02.297 "uuid": "60a8f92e-12f4-4968-8980-adbb931624c6", 00:30:02.297 "assigned_rate_limits": { 00:30:02.297 "rw_ios_per_sec": 0, 00:30:02.297 "rw_mbytes_per_sec": 0, 00:30:02.297 "r_mbytes_per_sec": 0, 00:30:02.297 "w_mbytes_per_sec": 0 00:30:02.297 }, 00:30:02.297 "claimed": false, 00:30:02.297 "zoned": false, 00:30:02.297 "supported_io_types": { 00:30:02.297 "read": true, 00:30:02.297 "write": true, 00:30:02.297 "unmap": true, 00:30:02.297 "flush": true, 00:30:02.297 "reset": true, 00:30:02.297 "nvme_admin": false, 00:30:02.297 "nvme_io": false, 00:30:02.297 "nvme_io_md": false, 00:30:02.297 "write_zeroes": true, 00:30:02.297 "zcopy": true, 00:30:02.297 "get_zone_info": false, 00:30:02.297 "zone_management": false, 00:30:02.297 "zone_append": false, 00:30:02.297 "compare": false, 00:30:02.297 "compare_and_write": false, 00:30:02.297 "abort": true, 00:30:02.297 "seek_hole": false, 00:30:02.297 "seek_data": false, 00:30:02.297 "copy": true, 00:30:02.297 "nvme_iov_md": false 00:30:02.297 }, 00:30:02.297 "memory_domains": [ 00:30:02.297 { 00:30:02.297 "dma_device_id": "system", 00:30:02.297 "dma_device_type": 1 00:30:02.297 }, 00:30:02.297 { 00:30:02.297 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:02.297 "dma_device_type": 2 00:30:02.297 } 00:30:02.297 ], 00:30:02.297 "driver_specific": {} 00:30:02.297 } 00:30:02.297 ] 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:02.297 [2024-11-20 13:50:09.263575] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:30:02.297 [2024-11-20 13:50:09.263779] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:30:02.297 [2024-11-20 13:50:09.263845] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:30:02.297 [2024-11-20 13:50:09.265607] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:30:02.297 [2024-11-20 13:50:09.265742] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:02.297 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:02.297 "name": "Existed_Raid", 00:30:02.297 "uuid": "0439a214-8d90-4729-a7c9-47262a21337a", 00:30:02.297 "strip_size_kb": 0, 00:30:02.297 "state": "configuring", 00:30:02.297 "raid_level": "raid1", 00:30:02.297 "superblock": true, 00:30:02.297 "num_base_bdevs": 4, 00:30:02.297 "num_base_bdevs_discovered": 3, 00:30:02.297 "num_base_bdevs_operational": 4, 00:30:02.297 "base_bdevs_list": [ 00:30:02.297 { 00:30:02.297 "name": "BaseBdev1", 00:30:02.297 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:02.297 "is_configured": false, 00:30:02.297 "data_offset": 0, 00:30:02.297 "data_size": 0 00:30:02.297 }, 00:30:02.297 { 00:30:02.298 "name": "BaseBdev2", 00:30:02.298 "uuid": "5eed79cb-fe93-466b-ad51-e3cc15d577e3", 00:30:02.298 "is_configured": true, 00:30:02.298 "data_offset": 2048, 00:30:02.298 "data_size": 63488 00:30:02.298 }, 00:30:02.298 { 00:30:02.298 "name": "BaseBdev3", 00:30:02.298 "uuid": "2636ee53-fea7-4da4-ad0e-32e4f6345fd3", 00:30:02.298 "is_configured": true, 00:30:02.298 "data_offset": 2048, 00:30:02.298 "data_size": 63488 00:30:02.298 }, 00:30:02.298 { 00:30:02.298 "name": "BaseBdev4", 00:30:02.298 "uuid": "60a8f92e-12f4-4968-8980-adbb931624c6", 00:30:02.298 "is_configured": true, 00:30:02.298 "data_offset": 2048, 00:30:02.298 "data_size": 63488 00:30:02.298 } 00:30:02.298 ] 00:30:02.298 }' 00:30:02.298 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:02.298 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:02.555 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:30:02.555 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:02.555 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:02.555 [2024-11-20 13:50:09.579655] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:30:02.555 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:02.555 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:30:02.555 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:02.555 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:02.555 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:02.555 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:02.555 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:02.555 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:02.555 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:02.555 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:02.555 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:02.555 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:02.555 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:02.555 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:02.555 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:02.555 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:02.875 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:02.875 "name": "Existed_Raid", 00:30:02.875 "uuid": "0439a214-8d90-4729-a7c9-47262a21337a", 00:30:02.875 "strip_size_kb": 0, 00:30:02.875 "state": "configuring", 00:30:02.875 "raid_level": "raid1", 00:30:02.875 "superblock": true, 00:30:02.875 "num_base_bdevs": 4, 00:30:02.875 "num_base_bdevs_discovered": 2, 00:30:02.875 "num_base_bdevs_operational": 4, 00:30:02.875 "base_bdevs_list": [ 00:30:02.875 { 00:30:02.875 "name": "BaseBdev1", 00:30:02.875 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:02.875 "is_configured": false, 00:30:02.875 "data_offset": 0, 00:30:02.875 "data_size": 0 00:30:02.875 }, 00:30:02.875 { 00:30:02.875 "name": null, 00:30:02.875 "uuid": "5eed79cb-fe93-466b-ad51-e3cc15d577e3", 00:30:02.875 "is_configured": false, 00:30:02.875 "data_offset": 0, 00:30:02.875 "data_size": 63488 00:30:02.875 }, 00:30:02.875 { 00:30:02.875 "name": "BaseBdev3", 00:30:02.875 "uuid": "2636ee53-fea7-4da4-ad0e-32e4f6345fd3", 00:30:02.875 "is_configured": true, 00:30:02.875 "data_offset": 2048, 00:30:02.875 "data_size": 63488 00:30:02.875 }, 00:30:02.875 { 00:30:02.875 "name": "BaseBdev4", 00:30:02.875 "uuid": "60a8f92e-12f4-4968-8980-adbb931624c6", 00:30:02.875 "is_configured": true, 00:30:02.875 "data_offset": 2048, 00:30:02.875 "data_size": 63488 00:30:02.875 } 00:30:02.875 ] 00:30:02.875 }' 00:30:02.875 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:02.875 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:02.875 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:02.876 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:02.876 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:02.876 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:30:02.876 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:02.876 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:30:02.876 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:30:02.876 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:02.876 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:03.134 [2024-11-20 13:50:09.936348] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:30:03.134 BaseBdev1 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:03.134 [ 00:30:03.134 { 00:30:03.134 "name": "BaseBdev1", 00:30:03.134 "aliases": [ 00:30:03.134 "8e252f28-7ec4-4ccd-b70c-b4f16b7eb383" 00:30:03.134 ], 00:30:03.134 "product_name": "Malloc disk", 00:30:03.134 "block_size": 512, 00:30:03.134 "num_blocks": 65536, 00:30:03.134 "uuid": "8e252f28-7ec4-4ccd-b70c-b4f16b7eb383", 00:30:03.134 "assigned_rate_limits": { 00:30:03.134 "rw_ios_per_sec": 0, 00:30:03.134 "rw_mbytes_per_sec": 0, 00:30:03.134 "r_mbytes_per_sec": 0, 00:30:03.134 "w_mbytes_per_sec": 0 00:30:03.134 }, 00:30:03.134 "claimed": true, 00:30:03.134 "claim_type": "exclusive_write", 00:30:03.134 "zoned": false, 00:30:03.134 "supported_io_types": { 00:30:03.134 "read": true, 00:30:03.134 "write": true, 00:30:03.134 "unmap": true, 00:30:03.134 "flush": true, 00:30:03.134 "reset": true, 00:30:03.134 "nvme_admin": false, 00:30:03.134 "nvme_io": false, 00:30:03.134 "nvme_io_md": false, 00:30:03.134 "write_zeroes": true, 00:30:03.134 "zcopy": true, 00:30:03.134 "get_zone_info": false, 00:30:03.134 "zone_management": false, 00:30:03.134 "zone_append": false, 00:30:03.134 "compare": false, 00:30:03.134 "compare_and_write": false, 00:30:03.134 "abort": true, 00:30:03.134 "seek_hole": false, 00:30:03.134 "seek_data": false, 00:30:03.134 "copy": true, 00:30:03.134 "nvme_iov_md": false 00:30:03.134 }, 00:30:03.134 "memory_domains": [ 00:30:03.134 { 00:30:03.134 "dma_device_id": "system", 00:30:03.134 "dma_device_type": 1 00:30:03.134 }, 00:30:03.134 { 00:30:03.134 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:03.134 "dma_device_type": 2 00:30:03.134 } 00:30:03.134 ], 00:30:03.134 "driver_specific": {} 00:30:03.134 } 00:30:03.134 ] 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:03.134 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:03.134 "name": "Existed_Raid", 00:30:03.134 "uuid": "0439a214-8d90-4729-a7c9-47262a21337a", 00:30:03.134 "strip_size_kb": 0, 00:30:03.134 "state": "configuring", 00:30:03.134 "raid_level": "raid1", 00:30:03.134 "superblock": true, 00:30:03.134 "num_base_bdevs": 4, 00:30:03.134 "num_base_bdevs_discovered": 3, 00:30:03.134 "num_base_bdevs_operational": 4, 00:30:03.134 "base_bdevs_list": [ 00:30:03.134 { 00:30:03.134 "name": "BaseBdev1", 00:30:03.134 "uuid": "8e252f28-7ec4-4ccd-b70c-b4f16b7eb383", 00:30:03.134 "is_configured": true, 00:30:03.134 "data_offset": 2048, 00:30:03.134 "data_size": 63488 00:30:03.134 }, 00:30:03.134 { 00:30:03.134 "name": null, 00:30:03.134 "uuid": "5eed79cb-fe93-466b-ad51-e3cc15d577e3", 00:30:03.135 "is_configured": false, 00:30:03.135 "data_offset": 0, 00:30:03.135 "data_size": 63488 00:30:03.135 }, 00:30:03.135 { 00:30:03.135 "name": "BaseBdev3", 00:30:03.135 "uuid": "2636ee53-fea7-4da4-ad0e-32e4f6345fd3", 00:30:03.135 "is_configured": true, 00:30:03.135 "data_offset": 2048, 00:30:03.135 "data_size": 63488 00:30:03.135 }, 00:30:03.135 { 00:30:03.135 "name": "BaseBdev4", 00:30:03.135 "uuid": "60a8f92e-12f4-4968-8980-adbb931624c6", 00:30:03.135 "is_configured": true, 00:30:03.135 "data_offset": 2048, 00:30:03.135 "data_size": 63488 00:30:03.135 } 00:30:03.135 ] 00:30:03.135 }' 00:30:03.135 13:50:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:03.135 13:50:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:03.393 [2024-11-20 13:50:10.312519] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:03.393 "name": "Existed_Raid", 00:30:03.393 "uuid": "0439a214-8d90-4729-a7c9-47262a21337a", 00:30:03.393 "strip_size_kb": 0, 00:30:03.393 "state": "configuring", 00:30:03.393 "raid_level": "raid1", 00:30:03.393 "superblock": true, 00:30:03.393 "num_base_bdevs": 4, 00:30:03.393 "num_base_bdevs_discovered": 2, 00:30:03.393 "num_base_bdevs_operational": 4, 00:30:03.393 "base_bdevs_list": [ 00:30:03.393 { 00:30:03.393 "name": "BaseBdev1", 00:30:03.393 "uuid": "8e252f28-7ec4-4ccd-b70c-b4f16b7eb383", 00:30:03.393 "is_configured": true, 00:30:03.393 "data_offset": 2048, 00:30:03.393 "data_size": 63488 00:30:03.393 }, 00:30:03.393 { 00:30:03.393 "name": null, 00:30:03.393 "uuid": "5eed79cb-fe93-466b-ad51-e3cc15d577e3", 00:30:03.393 "is_configured": false, 00:30:03.393 "data_offset": 0, 00:30:03.393 "data_size": 63488 00:30:03.393 }, 00:30:03.393 { 00:30:03.393 "name": null, 00:30:03.393 "uuid": "2636ee53-fea7-4da4-ad0e-32e4f6345fd3", 00:30:03.393 "is_configured": false, 00:30:03.393 "data_offset": 0, 00:30:03.393 "data_size": 63488 00:30:03.393 }, 00:30:03.393 { 00:30:03.393 "name": "BaseBdev4", 00:30:03.393 "uuid": "60a8f92e-12f4-4968-8980-adbb931624c6", 00:30:03.393 "is_configured": true, 00:30:03.393 "data_offset": 2048, 00:30:03.393 "data_size": 63488 00:30:03.393 } 00:30:03.393 ] 00:30:03.393 }' 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:03.393 13:50:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:03.652 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:30:03.652 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:03.652 13:50:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:03.652 13:50:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:03.652 13:50:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:03.652 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:30:03.652 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:30:03.652 13:50:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:03.652 13:50:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:03.652 [2024-11-20 13:50:10.664558] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:30:03.652 13:50:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:03.652 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:30:03.652 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:03.652 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:03.652 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:03.652 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:03.652 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:03.652 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:03.652 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:03.652 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:03.652 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:03.652 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:03.652 13:50:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:03.652 13:50:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:03.652 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:03.652 13:50:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:03.911 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:03.911 "name": "Existed_Raid", 00:30:03.911 "uuid": "0439a214-8d90-4729-a7c9-47262a21337a", 00:30:03.911 "strip_size_kb": 0, 00:30:03.911 "state": "configuring", 00:30:03.911 "raid_level": "raid1", 00:30:03.911 "superblock": true, 00:30:03.911 "num_base_bdevs": 4, 00:30:03.911 "num_base_bdevs_discovered": 3, 00:30:03.911 "num_base_bdevs_operational": 4, 00:30:03.911 "base_bdevs_list": [ 00:30:03.911 { 00:30:03.911 "name": "BaseBdev1", 00:30:03.911 "uuid": "8e252f28-7ec4-4ccd-b70c-b4f16b7eb383", 00:30:03.911 "is_configured": true, 00:30:03.911 "data_offset": 2048, 00:30:03.911 "data_size": 63488 00:30:03.911 }, 00:30:03.911 { 00:30:03.911 "name": null, 00:30:03.911 "uuid": "5eed79cb-fe93-466b-ad51-e3cc15d577e3", 00:30:03.911 "is_configured": false, 00:30:03.911 "data_offset": 0, 00:30:03.911 "data_size": 63488 00:30:03.911 }, 00:30:03.911 { 00:30:03.911 "name": "BaseBdev3", 00:30:03.911 "uuid": "2636ee53-fea7-4da4-ad0e-32e4f6345fd3", 00:30:03.911 "is_configured": true, 00:30:03.911 "data_offset": 2048, 00:30:03.911 "data_size": 63488 00:30:03.911 }, 00:30:03.911 { 00:30:03.911 "name": "BaseBdev4", 00:30:03.911 "uuid": "60a8f92e-12f4-4968-8980-adbb931624c6", 00:30:03.911 "is_configured": true, 00:30:03.911 "data_offset": 2048, 00:30:03.911 "data_size": 63488 00:30:03.911 } 00:30:03.911 ] 00:30:03.911 }' 00:30:03.911 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:03.911 13:50:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:04.169 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:04.169 13:50:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:04.169 13:50:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:04.169 13:50:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:30:04.169 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:04.169 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:30:04.169 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:30:04.169 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:04.169 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:04.169 [2024-11-20 13:50:11.032681] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:30:04.169 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:04.169 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:30:04.169 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:04.169 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:04.169 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:04.169 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:04.170 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:04.170 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:04.170 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:04.170 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:04.170 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:04.170 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:04.170 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:04.170 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:04.170 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:04.170 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:04.170 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:04.170 "name": "Existed_Raid", 00:30:04.170 "uuid": "0439a214-8d90-4729-a7c9-47262a21337a", 00:30:04.170 "strip_size_kb": 0, 00:30:04.170 "state": "configuring", 00:30:04.170 "raid_level": "raid1", 00:30:04.170 "superblock": true, 00:30:04.170 "num_base_bdevs": 4, 00:30:04.170 "num_base_bdevs_discovered": 2, 00:30:04.170 "num_base_bdevs_operational": 4, 00:30:04.170 "base_bdevs_list": [ 00:30:04.170 { 00:30:04.170 "name": null, 00:30:04.170 "uuid": "8e252f28-7ec4-4ccd-b70c-b4f16b7eb383", 00:30:04.170 "is_configured": false, 00:30:04.170 "data_offset": 0, 00:30:04.170 "data_size": 63488 00:30:04.170 }, 00:30:04.170 { 00:30:04.170 "name": null, 00:30:04.170 "uuid": "5eed79cb-fe93-466b-ad51-e3cc15d577e3", 00:30:04.170 "is_configured": false, 00:30:04.170 "data_offset": 0, 00:30:04.170 "data_size": 63488 00:30:04.170 }, 00:30:04.170 { 00:30:04.170 "name": "BaseBdev3", 00:30:04.170 "uuid": "2636ee53-fea7-4da4-ad0e-32e4f6345fd3", 00:30:04.170 "is_configured": true, 00:30:04.170 "data_offset": 2048, 00:30:04.170 "data_size": 63488 00:30:04.170 }, 00:30:04.170 { 00:30:04.170 "name": "BaseBdev4", 00:30:04.170 "uuid": "60a8f92e-12f4-4968-8980-adbb931624c6", 00:30:04.170 "is_configured": true, 00:30:04.170 "data_offset": 2048, 00:30:04.170 "data_size": 63488 00:30:04.170 } 00:30:04.170 ] 00:30:04.170 }' 00:30:04.170 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:04.170 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:04.428 [2024-11-20 13:50:11.443191] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:04.428 "name": "Existed_Raid", 00:30:04.428 "uuid": "0439a214-8d90-4729-a7c9-47262a21337a", 00:30:04.428 "strip_size_kb": 0, 00:30:04.428 "state": "configuring", 00:30:04.428 "raid_level": "raid1", 00:30:04.428 "superblock": true, 00:30:04.428 "num_base_bdevs": 4, 00:30:04.428 "num_base_bdevs_discovered": 3, 00:30:04.428 "num_base_bdevs_operational": 4, 00:30:04.428 "base_bdevs_list": [ 00:30:04.428 { 00:30:04.428 "name": null, 00:30:04.428 "uuid": "8e252f28-7ec4-4ccd-b70c-b4f16b7eb383", 00:30:04.428 "is_configured": false, 00:30:04.428 "data_offset": 0, 00:30:04.428 "data_size": 63488 00:30:04.428 }, 00:30:04.428 { 00:30:04.428 "name": "BaseBdev2", 00:30:04.428 "uuid": "5eed79cb-fe93-466b-ad51-e3cc15d577e3", 00:30:04.428 "is_configured": true, 00:30:04.428 "data_offset": 2048, 00:30:04.428 "data_size": 63488 00:30:04.428 }, 00:30:04.428 { 00:30:04.428 "name": "BaseBdev3", 00:30:04.428 "uuid": "2636ee53-fea7-4da4-ad0e-32e4f6345fd3", 00:30:04.428 "is_configured": true, 00:30:04.428 "data_offset": 2048, 00:30:04.428 "data_size": 63488 00:30:04.428 }, 00:30:04.428 { 00:30:04.428 "name": "BaseBdev4", 00:30:04.428 "uuid": "60a8f92e-12f4-4968-8980-adbb931624c6", 00:30:04.428 "is_configured": true, 00:30:04.428 "data_offset": 2048, 00:30:04.428 "data_size": 63488 00:30:04.428 } 00:30:04.428 ] 00:30:04.428 }' 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:04.428 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:04.995 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:04.995 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 8e252f28-7ec4-4ccd-b70c-b4f16b7eb383 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:04.996 [2024-11-20 13:50:11.828438] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:30:04.996 [2024-11-20 13:50:11.828635] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:30:04.996 [2024-11-20 13:50:11.828649] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:30:04.996 NewBaseBdev 00:30:04.996 [2024-11-20 13:50:11.828875] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:30:04.996 [2024-11-20 13:50:11.829016] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:30:04.996 [2024-11-20 13:50:11.829030] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:30:04.996 [2024-11-20 13:50:11.829145] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:04.996 [ 00:30:04.996 { 00:30:04.996 "name": "NewBaseBdev", 00:30:04.996 "aliases": [ 00:30:04.996 "8e252f28-7ec4-4ccd-b70c-b4f16b7eb383" 00:30:04.996 ], 00:30:04.996 "product_name": "Malloc disk", 00:30:04.996 "block_size": 512, 00:30:04.996 "num_blocks": 65536, 00:30:04.996 "uuid": "8e252f28-7ec4-4ccd-b70c-b4f16b7eb383", 00:30:04.996 "assigned_rate_limits": { 00:30:04.996 "rw_ios_per_sec": 0, 00:30:04.996 "rw_mbytes_per_sec": 0, 00:30:04.996 "r_mbytes_per_sec": 0, 00:30:04.996 "w_mbytes_per_sec": 0 00:30:04.996 }, 00:30:04.996 "claimed": true, 00:30:04.996 "claim_type": "exclusive_write", 00:30:04.996 "zoned": false, 00:30:04.996 "supported_io_types": { 00:30:04.996 "read": true, 00:30:04.996 "write": true, 00:30:04.996 "unmap": true, 00:30:04.996 "flush": true, 00:30:04.996 "reset": true, 00:30:04.996 "nvme_admin": false, 00:30:04.996 "nvme_io": false, 00:30:04.996 "nvme_io_md": false, 00:30:04.996 "write_zeroes": true, 00:30:04.996 "zcopy": true, 00:30:04.996 "get_zone_info": false, 00:30:04.996 "zone_management": false, 00:30:04.996 "zone_append": false, 00:30:04.996 "compare": false, 00:30:04.996 "compare_and_write": false, 00:30:04.996 "abort": true, 00:30:04.996 "seek_hole": false, 00:30:04.996 "seek_data": false, 00:30:04.996 "copy": true, 00:30:04.996 "nvme_iov_md": false 00:30:04.996 }, 00:30:04.996 "memory_domains": [ 00:30:04.996 { 00:30:04.996 "dma_device_id": "system", 00:30:04.996 "dma_device_type": 1 00:30:04.996 }, 00:30:04.996 { 00:30:04.996 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:04.996 "dma_device_type": 2 00:30:04.996 } 00:30:04.996 ], 00:30:04.996 "driver_specific": {} 00:30:04.996 } 00:30:04.996 ] 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:04.996 "name": "Existed_Raid", 00:30:04.996 "uuid": "0439a214-8d90-4729-a7c9-47262a21337a", 00:30:04.996 "strip_size_kb": 0, 00:30:04.996 "state": "online", 00:30:04.996 "raid_level": "raid1", 00:30:04.996 "superblock": true, 00:30:04.996 "num_base_bdevs": 4, 00:30:04.996 "num_base_bdevs_discovered": 4, 00:30:04.996 "num_base_bdevs_operational": 4, 00:30:04.996 "base_bdevs_list": [ 00:30:04.996 { 00:30:04.996 "name": "NewBaseBdev", 00:30:04.996 "uuid": "8e252f28-7ec4-4ccd-b70c-b4f16b7eb383", 00:30:04.996 "is_configured": true, 00:30:04.996 "data_offset": 2048, 00:30:04.996 "data_size": 63488 00:30:04.996 }, 00:30:04.996 { 00:30:04.996 "name": "BaseBdev2", 00:30:04.996 "uuid": "5eed79cb-fe93-466b-ad51-e3cc15d577e3", 00:30:04.996 "is_configured": true, 00:30:04.996 "data_offset": 2048, 00:30:04.996 "data_size": 63488 00:30:04.996 }, 00:30:04.996 { 00:30:04.996 "name": "BaseBdev3", 00:30:04.996 "uuid": "2636ee53-fea7-4da4-ad0e-32e4f6345fd3", 00:30:04.996 "is_configured": true, 00:30:04.996 "data_offset": 2048, 00:30:04.996 "data_size": 63488 00:30:04.996 }, 00:30:04.996 { 00:30:04.996 "name": "BaseBdev4", 00:30:04.996 "uuid": "60a8f92e-12f4-4968-8980-adbb931624c6", 00:30:04.996 "is_configured": true, 00:30:04.996 "data_offset": 2048, 00:30:04.996 "data_size": 63488 00:30:04.996 } 00:30:04.996 ] 00:30:04.996 }' 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:04.996 13:50:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:05.255 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:30:05.255 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:30:05.255 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:30:05.255 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:30:05.255 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:30:05.255 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:30:05.255 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:30:05.255 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:05.255 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:05.255 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:30:05.255 [2024-11-20 13:50:12.184912] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:30:05.255 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:05.255 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:30:05.255 "name": "Existed_Raid", 00:30:05.255 "aliases": [ 00:30:05.255 "0439a214-8d90-4729-a7c9-47262a21337a" 00:30:05.255 ], 00:30:05.255 "product_name": "Raid Volume", 00:30:05.255 "block_size": 512, 00:30:05.255 "num_blocks": 63488, 00:30:05.255 "uuid": "0439a214-8d90-4729-a7c9-47262a21337a", 00:30:05.255 "assigned_rate_limits": { 00:30:05.255 "rw_ios_per_sec": 0, 00:30:05.255 "rw_mbytes_per_sec": 0, 00:30:05.255 "r_mbytes_per_sec": 0, 00:30:05.255 "w_mbytes_per_sec": 0 00:30:05.255 }, 00:30:05.255 "claimed": false, 00:30:05.255 "zoned": false, 00:30:05.255 "supported_io_types": { 00:30:05.255 "read": true, 00:30:05.255 "write": true, 00:30:05.255 "unmap": false, 00:30:05.255 "flush": false, 00:30:05.255 "reset": true, 00:30:05.255 "nvme_admin": false, 00:30:05.255 "nvme_io": false, 00:30:05.255 "nvme_io_md": false, 00:30:05.256 "write_zeroes": true, 00:30:05.256 "zcopy": false, 00:30:05.256 "get_zone_info": false, 00:30:05.256 "zone_management": false, 00:30:05.256 "zone_append": false, 00:30:05.256 "compare": false, 00:30:05.256 "compare_and_write": false, 00:30:05.256 "abort": false, 00:30:05.256 "seek_hole": false, 00:30:05.256 "seek_data": false, 00:30:05.256 "copy": false, 00:30:05.256 "nvme_iov_md": false 00:30:05.256 }, 00:30:05.256 "memory_domains": [ 00:30:05.256 { 00:30:05.256 "dma_device_id": "system", 00:30:05.256 "dma_device_type": 1 00:30:05.256 }, 00:30:05.256 { 00:30:05.256 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:05.256 "dma_device_type": 2 00:30:05.256 }, 00:30:05.256 { 00:30:05.256 "dma_device_id": "system", 00:30:05.256 "dma_device_type": 1 00:30:05.256 }, 00:30:05.256 { 00:30:05.256 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:05.256 "dma_device_type": 2 00:30:05.256 }, 00:30:05.256 { 00:30:05.256 "dma_device_id": "system", 00:30:05.256 "dma_device_type": 1 00:30:05.256 }, 00:30:05.256 { 00:30:05.256 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:05.256 "dma_device_type": 2 00:30:05.256 }, 00:30:05.256 { 00:30:05.256 "dma_device_id": "system", 00:30:05.256 "dma_device_type": 1 00:30:05.256 }, 00:30:05.256 { 00:30:05.256 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:05.256 "dma_device_type": 2 00:30:05.256 } 00:30:05.256 ], 00:30:05.256 "driver_specific": { 00:30:05.256 "raid": { 00:30:05.256 "uuid": "0439a214-8d90-4729-a7c9-47262a21337a", 00:30:05.256 "strip_size_kb": 0, 00:30:05.256 "state": "online", 00:30:05.256 "raid_level": "raid1", 00:30:05.256 "superblock": true, 00:30:05.256 "num_base_bdevs": 4, 00:30:05.256 "num_base_bdevs_discovered": 4, 00:30:05.256 "num_base_bdevs_operational": 4, 00:30:05.256 "base_bdevs_list": [ 00:30:05.256 { 00:30:05.256 "name": "NewBaseBdev", 00:30:05.256 "uuid": "8e252f28-7ec4-4ccd-b70c-b4f16b7eb383", 00:30:05.256 "is_configured": true, 00:30:05.256 "data_offset": 2048, 00:30:05.256 "data_size": 63488 00:30:05.256 }, 00:30:05.256 { 00:30:05.256 "name": "BaseBdev2", 00:30:05.256 "uuid": "5eed79cb-fe93-466b-ad51-e3cc15d577e3", 00:30:05.256 "is_configured": true, 00:30:05.256 "data_offset": 2048, 00:30:05.256 "data_size": 63488 00:30:05.256 }, 00:30:05.256 { 00:30:05.256 "name": "BaseBdev3", 00:30:05.256 "uuid": "2636ee53-fea7-4da4-ad0e-32e4f6345fd3", 00:30:05.256 "is_configured": true, 00:30:05.256 "data_offset": 2048, 00:30:05.256 "data_size": 63488 00:30:05.256 }, 00:30:05.256 { 00:30:05.256 "name": "BaseBdev4", 00:30:05.256 "uuid": "60a8f92e-12f4-4968-8980-adbb931624c6", 00:30:05.256 "is_configured": true, 00:30:05.256 "data_offset": 2048, 00:30:05.256 "data_size": 63488 00:30:05.256 } 00:30:05.256 ] 00:30:05.256 } 00:30:05.256 } 00:30:05.256 }' 00:30:05.256 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:30:05.256 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:30:05.256 BaseBdev2 00:30:05.256 BaseBdev3 00:30:05.256 BaseBdev4' 00:30:05.256 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:05.256 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:30:05.256 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:05.256 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:30:05.256 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:05.256 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:05.256 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:05.256 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:05.514 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:05.514 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:05.514 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:05.514 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:30:05.514 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:05.514 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:05.514 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:05.514 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:05.514 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:05.514 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:05.514 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:05.514 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:30:05.514 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:05.514 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:05.514 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:05.514 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:05.514 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:05.514 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:05.514 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:05.514 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:30:05.514 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:05.514 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:05.514 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:05.514 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:05.515 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:05.515 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:05.515 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:30:05.515 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:05.515 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:05.515 [2024-11-20 13:50:12.412616] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:30:05.515 [2024-11-20 13:50:12.412655] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:30:05.515 [2024-11-20 13:50:12.412735] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:30:05.515 [2024-11-20 13:50:12.413022] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:30:05.515 [2024-11-20 13:50:12.413035] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:30:05.515 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:05.515 13:50:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 71834 00:30:05.515 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 71834 ']' 00:30:05.515 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 71834 00:30:05.515 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:30:05.515 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:30:05.515 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71834 00:30:05.515 killing process with pid 71834 00:30:05.515 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:30:05.515 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:30:05.515 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71834' 00:30:05.515 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 71834 00:30:05.515 [2024-11-20 13:50:12.444115] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:30:05.515 13:50:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 71834 00:30:05.774 [2024-11-20 13:50:12.669677] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:30:06.414 13:50:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:30:06.414 00:30:06.414 real 0m8.185s 00:30:06.414 user 0m13.005s 00:30:06.414 sys 0m1.465s 00:30:06.414 13:50:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:30:06.414 ************************************ 00:30:06.414 END TEST raid_state_function_test_sb 00:30:06.414 ************************************ 00:30:06.414 13:50:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:06.414 13:50:13 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 4 00:30:06.414 13:50:13 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:30:06.414 13:50:13 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:30:06.414 13:50:13 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:30:06.414 ************************************ 00:30:06.414 START TEST raid_superblock_test 00:30:06.414 ************************************ 00:30:06.414 13:50:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 4 00:30:06.414 13:50:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:30:06.414 13:50:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:30:06.414 13:50:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:30:06.414 13:50:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:30:06.414 13:50:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:30:06.414 13:50:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:30:06.414 13:50:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:30:06.414 13:50:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:30:06.414 13:50:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:30:06.414 13:50:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:30:06.414 13:50:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:30:06.414 13:50:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:30:06.414 13:50:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:30:06.414 13:50:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:30:06.414 13:50:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:30:06.414 13:50:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=72470 00:30:06.414 13:50:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 72470 00:30:06.414 13:50:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 72470 ']' 00:30:06.414 13:50:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:30:06.414 13:50:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:30:06.414 13:50:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:30:06.414 13:50:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:30:06.414 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:30:06.414 13:50:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:30:06.414 13:50:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:06.414 [2024-11-20 13:50:13.434432] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:30:06.414 [2024-11-20 13:50:13.434702] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72470 ] 00:30:06.672 [2024-11-20 13:50:13.598927] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:30:06.672 [2024-11-20 13:50:13.718602] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:30:06.937 [2024-11-20 13:50:13.867669] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:30:06.937 [2024-11-20 13:50:13.867879] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:30:07.502 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:30:07.502 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:30:07.502 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:30:07.502 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:30:07.502 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:30:07.502 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:30:07.502 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:30:07.502 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:30:07.502 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:07.503 malloc1 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:07.503 [2024-11-20 13:50:14.392641] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:30:07.503 [2024-11-20 13:50:14.392876] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:07.503 [2024-11-20 13:50:14.392907] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:30:07.503 [2024-11-20 13:50:14.392918] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:07.503 [2024-11-20 13:50:14.395262] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:07.503 [2024-11-20 13:50:14.395297] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:30:07.503 pt1 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:07.503 malloc2 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:07.503 [2024-11-20 13:50:14.431110] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:30:07.503 [2024-11-20 13:50:14.431325] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:07.503 [2024-11-20 13:50:14.431359] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:30:07.503 [2024-11-20 13:50:14.431369] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:07.503 [2024-11-20 13:50:14.433615] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:07.503 [2024-11-20 13:50:14.433650] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:30:07.503 pt2 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:07.503 malloc3 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:07.503 [2024-11-20 13:50:14.482826] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:30:07.503 [2024-11-20 13:50:14.483058] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:07.503 [2024-11-20 13:50:14.483091] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:30:07.503 [2024-11-20 13:50:14.483102] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:07.503 [2024-11-20 13:50:14.485348] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:07.503 [2024-11-20 13:50:14.485384] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:30:07.503 pt3 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:07.503 malloc4 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:07.503 [2024-11-20 13:50:14.521168] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:30:07.503 [2024-11-20 13:50:14.521223] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:07.503 [2024-11-20 13:50:14.521241] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:30:07.503 [2024-11-20 13:50:14.521250] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:07.503 [2024-11-20 13:50:14.523460] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:07.503 [2024-11-20 13:50:14.523493] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:30:07.503 pt4 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:30:07.503 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:30:07.504 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:30:07.504 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:07.504 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:07.504 [2024-11-20 13:50:14.529202] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:30:07.504 [2024-11-20 13:50:14.531160] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:30:07.504 [2024-11-20 13:50:14.531224] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:30:07.504 [2024-11-20 13:50:14.531291] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:30:07.504 [2024-11-20 13:50:14.531478] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:30:07.504 [2024-11-20 13:50:14.531493] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:30:07.504 [2024-11-20 13:50:14.531756] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:30:07.504 [2024-11-20 13:50:14.531918] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:30:07.504 [2024-11-20 13:50:14.531931] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:30:07.504 [2024-11-20 13:50:14.532091] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:07.504 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:07.504 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:30:07.504 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:07.504 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:07.504 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:07.504 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:07.504 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:07.504 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:07.504 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:07.504 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:07.504 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:07.504 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:07.504 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:07.504 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:07.504 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:07.504 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:07.762 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:07.762 "name": "raid_bdev1", 00:30:07.762 "uuid": "3524d2e1-6a61-4f7c-ad81-b96908001c6f", 00:30:07.762 "strip_size_kb": 0, 00:30:07.762 "state": "online", 00:30:07.762 "raid_level": "raid1", 00:30:07.762 "superblock": true, 00:30:07.762 "num_base_bdevs": 4, 00:30:07.762 "num_base_bdevs_discovered": 4, 00:30:07.762 "num_base_bdevs_operational": 4, 00:30:07.762 "base_bdevs_list": [ 00:30:07.763 { 00:30:07.763 "name": "pt1", 00:30:07.763 "uuid": "00000000-0000-0000-0000-000000000001", 00:30:07.763 "is_configured": true, 00:30:07.763 "data_offset": 2048, 00:30:07.763 "data_size": 63488 00:30:07.763 }, 00:30:07.763 { 00:30:07.763 "name": "pt2", 00:30:07.763 "uuid": "00000000-0000-0000-0000-000000000002", 00:30:07.763 "is_configured": true, 00:30:07.763 "data_offset": 2048, 00:30:07.763 "data_size": 63488 00:30:07.763 }, 00:30:07.763 { 00:30:07.763 "name": "pt3", 00:30:07.763 "uuid": "00000000-0000-0000-0000-000000000003", 00:30:07.763 "is_configured": true, 00:30:07.763 "data_offset": 2048, 00:30:07.763 "data_size": 63488 00:30:07.763 }, 00:30:07.763 { 00:30:07.763 "name": "pt4", 00:30:07.763 "uuid": "00000000-0000-0000-0000-000000000004", 00:30:07.763 "is_configured": true, 00:30:07.763 "data_offset": 2048, 00:30:07.763 "data_size": 63488 00:30:07.763 } 00:30:07.763 ] 00:30:07.763 }' 00:30:07.763 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:07.763 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:07.763 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:30:07.763 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:30:07.763 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:30:07.763 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:30:07.763 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:30:07.763 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:30:07.763 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:30:07.763 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:30:07.763 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:07.763 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:07.763 [2024-11-20 13:50:14.809644] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:30:08.022 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:08.022 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:30:08.022 "name": "raid_bdev1", 00:30:08.022 "aliases": [ 00:30:08.022 "3524d2e1-6a61-4f7c-ad81-b96908001c6f" 00:30:08.022 ], 00:30:08.022 "product_name": "Raid Volume", 00:30:08.022 "block_size": 512, 00:30:08.022 "num_blocks": 63488, 00:30:08.022 "uuid": "3524d2e1-6a61-4f7c-ad81-b96908001c6f", 00:30:08.022 "assigned_rate_limits": { 00:30:08.022 "rw_ios_per_sec": 0, 00:30:08.022 "rw_mbytes_per_sec": 0, 00:30:08.022 "r_mbytes_per_sec": 0, 00:30:08.022 "w_mbytes_per_sec": 0 00:30:08.022 }, 00:30:08.022 "claimed": false, 00:30:08.022 "zoned": false, 00:30:08.022 "supported_io_types": { 00:30:08.022 "read": true, 00:30:08.022 "write": true, 00:30:08.022 "unmap": false, 00:30:08.022 "flush": false, 00:30:08.022 "reset": true, 00:30:08.022 "nvme_admin": false, 00:30:08.022 "nvme_io": false, 00:30:08.022 "nvme_io_md": false, 00:30:08.022 "write_zeroes": true, 00:30:08.022 "zcopy": false, 00:30:08.022 "get_zone_info": false, 00:30:08.022 "zone_management": false, 00:30:08.022 "zone_append": false, 00:30:08.022 "compare": false, 00:30:08.022 "compare_and_write": false, 00:30:08.022 "abort": false, 00:30:08.022 "seek_hole": false, 00:30:08.022 "seek_data": false, 00:30:08.022 "copy": false, 00:30:08.022 "nvme_iov_md": false 00:30:08.022 }, 00:30:08.022 "memory_domains": [ 00:30:08.022 { 00:30:08.022 "dma_device_id": "system", 00:30:08.022 "dma_device_type": 1 00:30:08.022 }, 00:30:08.022 { 00:30:08.022 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:08.022 "dma_device_type": 2 00:30:08.022 }, 00:30:08.022 { 00:30:08.022 "dma_device_id": "system", 00:30:08.022 "dma_device_type": 1 00:30:08.022 }, 00:30:08.022 { 00:30:08.022 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:08.022 "dma_device_type": 2 00:30:08.022 }, 00:30:08.022 { 00:30:08.022 "dma_device_id": "system", 00:30:08.022 "dma_device_type": 1 00:30:08.022 }, 00:30:08.022 { 00:30:08.022 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:08.022 "dma_device_type": 2 00:30:08.022 }, 00:30:08.022 { 00:30:08.022 "dma_device_id": "system", 00:30:08.022 "dma_device_type": 1 00:30:08.022 }, 00:30:08.022 { 00:30:08.022 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:08.022 "dma_device_type": 2 00:30:08.022 } 00:30:08.022 ], 00:30:08.022 "driver_specific": { 00:30:08.022 "raid": { 00:30:08.022 "uuid": "3524d2e1-6a61-4f7c-ad81-b96908001c6f", 00:30:08.022 "strip_size_kb": 0, 00:30:08.022 "state": "online", 00:30:08.022 "raid_level": "raid1", 00:30:08.022 "superblock": true, 00:30:08.022 "num_base_bdevs": 4, 00:30:08.022 "num_base_bdevs_discovered": 4, 00:30:08.022 "num_base_bdevs_operational": 4, 00:30:08.022 "base_bdevs_list": [ 00:30:08.022 { 00:30:08.022 "name": "pt1", 00:30:08.022 "uuid": "00000000-0000-0000-0000-000000000001", 00:30:08.022 "is_configured": true, 00:30:08.022 "data_offset": 2048, 00:30:08.022 "data_size": 63488 00:30:08.022 }, 00:30:08.022 { 00:30:08.022 "name": "pt2", 00:30:08.022 "uuid": "00000000-0000-0000-0000-000000000002", 00:30:08.022 "is_configured": true, 00:30:08.022 "data_offset": 2048, 00:30:08.023 "data_size": 63488 00:30:08.023 }, 00:30:08.023 { 00:30:08.023 "name": "pt3", 00:30:08.023 "uuid": "00000000-0000-0000-0000-000000000003", 00:30:08.023 "is_configured": true, 00:30:08.023 "data_offset": 2048, 00:30:08.023 "data_size": 63488 00:30:08.023 }, 00:30:08.023 { 00:30:08.023 "name": "pt4", 00:30:08.023 "uuid": "00000000-0000-0000-0000-000000000004", 00:30:08.023 "is_configured": true, 00:30:08.023 "data_offset": 2048, 00:30:08.023 "data_size": 63488 00:30:08.023 } 00:30:08.023 ] 00:30:08.023 } 00:30:08.023 } 00:30:08.023 }' 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:30:08.023 pt2 00:30:08.023 pt3 00:30:08.023 pt4' 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:08.023 13:50:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:08.023 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:08.023 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:08.023 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:08.023 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:30:08.023 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:30:08.023 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:08.023 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:08.023 [2024-11-20 13:50:15.033643] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:30:08.023 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:08.023 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=3524d2e1-6a61-4f7c-ad81-b96908001c6f 00:30:08.023 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 3524d2e1-6a61-4f7c-ad81-b96908001c6f ']' 00:30:08.023 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:30:08.023 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:08.023 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:08.023 [2024-11-20 13:50:15.065316] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:30:08.023 [2024-11-20 13:50:15.065347] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:30:08.023 [2024-11-20 13:50:15.065437] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:30:08.023 [2024-11-20 13:50:15.065538] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:30:08.023 [2024-11-20 13:50:15.065554] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:30:08.023 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:08.023 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:08.023 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:08.023 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:08.023 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:30:08.023 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:08.281 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:08.281 [2024-11-20 13:50:15.169369] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:30:08.281 [2024-11-20 13:50:15.171417] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:30:08.281 [2024-11-20 13:50:15.171469] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:30:08.281 [2024-11-20 13:50:15.171506] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:30:08.281 [2024-11-20 13:50:15.171559] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:30:08.281 [2024-11-20 13:50:15.171615] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:30:08.281 [2024-11-20 13:50:15.171636] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:30:08.281 [2024-11-20 13:50:15.171656] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:30:08.281 [2024-11-20 13:50:15.171670] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:30:08.281 [2024-11-20 13:50:15.171681] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:30:08.281 request: 00:30:08.281 { 00:30:08.281 "name": "raid_bdev1", 00:30:08.281 "raid_level": "raid1", 00:30:08.281 "base_bdevs": [ 00:30:08.281 "malloc1", 00:30:08.281 "malloc2", 00:30:08.281 "malloc3", 00:30:08.281 "malloc4" 00:30:08.281 ], 00:30:08.281 "superblock": false, 00:30:08.281 "method": "bdev_raid_create", 00:30:08.281 "req_id": 1 00:30:08.281 } 00:30:08.282 Got JSON-RPC error response 00:30:08.282 response: 00:30:08.282 { 00:30:08.282 "code": -17, 00:30:08.282 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:30:08.282 } 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:08.282 [2024-11-20 13:50:15.213349] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:30:08.282 [2024-11-20 13:50:15.213417] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:08.282 [2024-11-20 13:50:15.213435] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:30:08.282 [2024-11-20 13:50:15.213446] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:08.282 [2024-11-20 13:50:15.215830] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:08.282 [2024-11-20 13:50:15.215873] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:30:08.282 [2024-11-20 13:50:15.215960] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:30:08.282 [2024-11-20 13:50:15.216033] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:30:08.282 pt1 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 4 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:08.282 "name": "raid_bdev1", 00:30:08.282 "uuid": "3524d2e1-6a61-4f7c-ad81-b96908001c6f", 00:30:08.282 "strip_size_kb": 0, 00:30:08.282 "state": "configuring", 00:30:08.282 "raid_level": "raid1", 00:30:08.282 "superblock": true, 00:30:08.282 "num_base_bdevs": 4, 00:30:08.282 "num_base_bdevs_discovered": 1, 00:30:08.282 "num_base_bdevs_operational": 4, 00:30:08.282 "base_bdevs_list": [ 00:30:08.282 { 00:30:08.282 "name": "pt1", 00:30:08.282 "uuid": "00000000-0000-0000-0000-000000000001", 00:30:08.282 "is_configured": true, 00:30:08.282 "data_offset": 2048, 00:30:08.282 "data_size": 63488 00:30:08.282 }, 00:30:08.282 { 00:30:08.282 "name": null, 00:30:08.282 "uuid": "00000000-0000-0000-0000-000000000002", 00:30:08.282 "is_configured": false, 00:30:08.282 "data_offset": 2048, 00:30:08.282 "data_size": 63488 00:30:08.282 }, 00:30:08.282 { 00:30:08.282 "name": null, 00:30:08.282 "uuid": "00000000-0000-0000-0000-000000000003", 00:30:08.282 "is_configured": false, 00:30:08.282 "data_offset": 2048, 00:30:08.282 "data_size": 63488 00:30:08.282 }, 00:30:08.282 { 00:30:08.282 "name": null, 00:30:08.282 "uuid": "00000000-0000-0000-0000-000000000004", 00:30:08.282 "is_configured": false, 00:30:08.282 "data_offset": 2048, 00:30:08.282 "data_size": 63488 00:30:08.282 } 00:30:08.282 ] 00:30:08.282 }' 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:08.282 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:08.539 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:30:08.539 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:30:08.539 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:08.539 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:08.539 [2024-11-20 13:50:15.533457] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:30:08.539 [2024-11-20 13:50:15.533535] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:08.539 [2024-11-20 13:50:15.533556] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:30:08.539 [2024-11-20 13:50:15.533568] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:08.539 [2024-11-20 13:50:15.534044] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:08.539 [2024-11-20 13:50:15.534063] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:30:08.539 [2024-11-20 13:50:15.534142] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:30:08.539 [2024-11-20 13:50:15.534166] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:30:08.539 pt2 00:30:08.539 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:08.539 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:30:08.539 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:08.539 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:08.539 [2024-11-20 13:50:15.541445] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:30:08.539 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:08.539 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 4 00:30:08.539 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:08.539 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:08.539 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:08.539 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:08.539 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:08.539 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:08.539 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:08.539 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:08.539 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:08.539 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:08.539 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:08.539 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:08.539 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:08.539 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:08.539 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:08.539 "name": "raid_bdev1", 00:30:08.539 "uuid": "3524d2e1-6a61-4f7c-ad81-b96908001c6f", 00:30:08.539 "strip_size_kb": 0, 00:30:08.539 "state": "configuring", 00:30:08.539 "raid_level": "raid1", 00:30:08.539 "superblock": true, 00:30:08.539 "num_base_bdevs": 4, 00:30:08.539 "num_base_bdevs_discovered": 1, 00:30:08.539 "num_base_bdevs_operational": 4, 00:30:08.539 "base_bdevs_list": [ 00:30:08.539 { 00:30:08.539 "name": "pt1", 00:30:08.539 "uuid": "00000000-0000-0000-0000-000000000001", 00:30:08.539 "is_configured": true, 00:30:08.539 "data_offset": 2048, 00:30:08.539 "data_size": 63488 00:30:08.539 }, 00:30:08.539 { 00:30:08.539 "name": null, 00:30:08.539 "uuid": "00000000-0000-0000-0000-000000000002", 00:30:08.539 "is_configured": false, 00:30:08.539 "data_offset": 0, 00:30:08.540 "data_size": 63488 00:30:08.540 }, 00:30:08.540 { 00:30:08.540 "name": null, 00:30:08.540 "uuid": "00000000-0000-0000-0000-000000000003", 00:30:08.540 "is_configured": false, 00:30:08.540 "data_offset": 2048, 00:30:08.540 "data_size": 63488 00:30:08.540 }, 00:30:08.540 { 00:30:08.540 "name": null, 00:30:08.540 "uuid": "00000000-0000-0000-0000-000000000004", 00:30:08.540 "is_configured": false, 00:30:08.540 "data_offset": 2048, 00:30:08.540 "data_size": 63488 00:30:08.540 } 00:30:08.540 ] 00:30:08.540 }' 00:30:08.540 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:08.540 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:09.104 [2024-11-20 13:50:15.877550] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:30:09.104 [2024-11-20 13:50:15.877626] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:09.104 [2024-11-20 13:50:15.877646] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:30:09.104 [2024-11-20 13:50:15.877656] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:09.104 [2024-11-20 13:50:15.878154] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:09.104 [2024-11-20 13:50:15.878176] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:30:09.104 [2024-11-20 13:50:15.878259] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:30:09.104 [2024-11-20 13:50:15.878280] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:30:09.104 pt2 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:09.104 [2024-11-20 13:50:15.885493] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:30:09.104 [2024-11-20 13:50:15.885544] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:09.104 [2024-11-20 13:50:15.885563] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:30:09.104 [2024-11-20 13:50:15.885571] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:09.104 [2024-11-20 13:50:15.885990] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:09.104 [2024-11-20 13:50:15.886008] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:30:09.104 [2024-11-20 13:50:15.886074] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:30:09.104 [2024-11-20 13:50:15.886091] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:30:09.104 pt3 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:09.104 [2024-11-20 13:50:15.893468] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:30:09.104 [2024-11-20 13:50:15.893509] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:09.104 [2024-11-20 13:50:15.893526] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:30:09.104 [2024-11-20 13:50:15.893535] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:09.104 [2024-11-20 13:50:15.893902] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:09.104 [2024-11-20 13:50:15.893927] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:30:09.104 [2024-11-20 13:50:15.894003] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:30:09.104 [2024-11-20 13:50:15.894024] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:30:09.104 [2024-11-20 13:50:15.894168] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:30:09.104 [2024-11-20 13:50:15.894178] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:30:09.104 [2024-11-20 13:50:15.894440] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:30:09.104 [2024-11-20 13:50:15.894589] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:30:09.104 [2024-11-20 13:50:15.894600] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:30:09.104 [2024-11-20 13:50:15.894728] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:09.104 pt4 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:09.104 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:09.104 "name": "raid_bdev1", 00:30:09.104 "uuid": "3524d2e1-6a61-4f7c-ad81-b96908001c6f", 00:30:09.104 "strip_size_kb": 0, 00:30:09.104 "state": "online", 00:30:09.104 "raid_level": "raid1", 00:30:09.104 "superblock": true, 00:30:09.104 "num_base_bdevs": 4, 00:30:09.104 "num_base_bdevs_discovered": 4, 00:30:09.104 "num_base_bdevs_operational": 4, 00:30:09.104 "base_bdevs_list": [ 00:30:09.104 { 00:30:09.104 "name": "pt1", 00:30:09.104 "uuid": "00000000-0000-0000-0000-000000000001", 00:30:09.104 "is_configured": true, 00:30:09.104 "data_offset": 2048, 00:30:09.104 "data_size": 63488 00:30:09.104 }, 00:30:09.104 { 00:30:09.104 "name": "pt2", 00:30:09.105 "uuid": "00000000-0000-0000-0000-000000000002", 00:30:09.105 "is_configured": true, 00:30:09.105 "data_offset": 2048, 00:30:09.105 "data_size": 63488 00:30:09.105 }, 00:30:09.105 { 00:30:09.105 "name": "pt3", 00:30:09.105 "uuid": "00000000-0000-0000-0000-000000000003", 00:30:09.105 "is_configured": true, 00:30:09.105 "data_offset": 2048, 00:30:09.105 "data_size": 63488 00:30:09.105 }, 00:30:09.105 { 00:30:09.105 "name": "pt4", 00:30:09.105 "uuid": "00000000-0000-0000-0000-000000000004", 00:30:09.105 "is_configured": true, 00:30:09.105 "data_offset": 2048, 00:30:09.105 "data_size": 63488 00:30:09.105 } 00:30:09.105 ] 00:30:09.105 }' 00:30:09.105 13:50:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:09.105 13:50:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:09.364 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:30:09.364 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:30:09.364 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:30:09.364 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:30:09.364 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:30:09.364 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:30:09.364 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:30:09.364 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:30:09.364 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:09.364 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:09.364 [2024-11-20 13:50:16.233974] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:30:09.364 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:09.364 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:30:09.364 "name": "raid_bdev1", 00:30:09.364 "aliases": [ 00:30:09.364 "3524d2e1-6a61-4f7c-ad81-b96908001c6f" 00:30:09.364 ], 00:30:09.364 "product_name": "Raid Volume", 00:30:09.364 "block_size": 512, 00:30:09.364 "num_blocks": 63488, 00:30:09.364 "uuid": "3524d2e1-6a61-4f7c-ad81-b96908001c6f", 00:30:09.364 "assigned_rate_limits": { 00:30:09.364 "rw_ios_per_sec": 0, 00:30:09.364 "rw_mbytes_per_sec": 0, 00:30:09.364 "r_mbytes_per_sec": 0, 00:30:09.364 "w_mbytes_per_sec": 0 00:30:09.364 }, 00:30:09.364 "claimed": false, 00:30:09.364 "zoned": false, 00:30:09.364 "supported_io_types": { 00:30:09.364 "read": true, 00:30:09.364 "write": true, 00:30:09.364 "unmap": false, 00:30:09.364 "flush": false, 00:30:09.364 "reset": true, 00:30:09.364 "nvme_admin": false, 00:30:09.364 "nvme_io": false, 00:30:09.364 "nvme_io_md": false, 00:30:09.364 "write_zeroes": true, 00:30:09.364 "zcopy": false, 00:30:09.364 "get_zone_info": false, 00:30:09.364 "zone_management": false, 00:30:09.364 "zone_append": false, 00:30:09.364 "compare": false, 00:30:09.364 "compare_and_write": false, 00:30:09.364 "abort": false, 00:30:09.364 "seek_hole": false, 00:30:09.364 "seek_data": false, 00:30:09.364 "copy": false, 00:30:09.364 "nvme_iov_md": false 00:30:09.364 }, 00:30:09.364 "memory_domains": [ 00:30:09.364 { 00:30:09.364 "dma_device_id": "system", 00:30:09.364 "dma_device_type": 1 00:30:09.364 }, 00:30:09.364 { 00:30:09.364 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:09.364 "dma_device_type": 2 00:30:09.364 }, 00:30:09.364 { 00:30:09.364 "dma_device_id": "system", 00:30:09.364 "dma_device_type": 1 00:30:09.364 }, 00:30:09.364 { 00:30:09.364 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:09.364 "dma_device_type": 2 00:30:09.364 }, 00:30:09.364 { 00:30:09.364 "dma_device_id": "system", 00:30:09.364 "dma_device_type": 1 00:30:09.364 }, 00:30:09.364 { 00:30:09.364 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:09.364 "dma_device_type": 2 00:30:09.364 }, 00:30:09.364 { 00:30:09.364 "dma_device_id": "system", 00:30:09.364 "dma_device_type": 1 00:30:09.364 }, 00:30:09.364 { 00:30:09.364 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:09.364 "dma_device_type": 2 00:30:09.364 } 00:30:09.364 ], 00:30:09.364 "driver_specific": { 00:30:09.364 "raid": { 00:30:09.364 "uuid": "3524d2e1-6a61-4f7c-ad81-b96908001c6f", 00:30:09.364 "strip_size_kb": 0, 00:30:09.364 "state": "online", 00:30:09.364 "raid_level": "raid1", 00:30:09.364 "superblock": true, 00:30:09.364 "num_base_bdevs": 4, 00:30:09.364 "num_base_bdevs_discovered": 4, 00:30:09.364 "num_base_bdevs_operational": 4, 00:30:09.364 "base_bdevs_list": [ 00:30:09.364 { 00:30:09.364 "name": "pt1", 00:30:09.364 "uuid": "00000000-0000-0000-0000-000000000001", 00:30:09.364 "is_configured": true, 00:30:09.364 "data_offset": 2048, 00:30:09.364 "data_size": 63488 00:30:09.364 }, 00:30:09.364 { 00:30:09.364 "name": "pt2", 00:30:09.364 "uuid": "00000000-0000-0000-0000-000000000002", 00:30:09.364 "is_configured": true, 00:30:09.364 "data_offset": 2048, 00:30:09.364 "data_size": 63488 00:30:09.364 }, 00:30:09.364 { 00:30:09.364 "name": "pt3", 00:30:09.364 "uuid": "00000000-0000-0000-0000-000000000003", 00:30:09.364 "is_configured": true, 00:30:09.364 "data_offset": 2048, 00:30:09.364 "data_size": 63488 00:30:09.364 }, 00:30:09.364 { 00:30:09.364 "name": "pt4", 00:30:09.364 "uuid": "00000000-0000-0000-0000-000000000004", 00:30:09.364 "is_configured": true, 00:30:09.364 "data_offset": 2048, 00:30:09.364 "data_size": 63488 00:30:09.364 } 00:30:09.364 ] 00:30:09.364 } 00:30:09.364 } 00:30:09.364 }' 00:30:09.364 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:30:09.364 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:30:09.364 pt2 00:30:09.364 pt3 00:30:09.364 pt4' 00:30:09.364 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:09.365 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:09.624 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:09.624 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:09.624 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:09.624 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:30:09.624 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:30:09.624 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:09.624 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:09.624 [2024-11-20 13:50:16.469975] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:30:09.624 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:09.624 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 3524d2e1-6a61-4f7c-ad81-b96908001c6f '!=' 3524d2e1-6a61-4f7c-ad81-b96908001c6f ']' 00:30:09.624 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:30:09.624 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:30:09.624 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:30:09.624 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:30:09.624 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:09.624 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:09.624 [2024-11-20 13:50:16.501706] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:30:09.624 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:09.624 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:30:09.624 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:09.624 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:09.624 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:09.624 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:09.624 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:30:09.625 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:09.625 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:09.625 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:09.625 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:09.625 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:09.625 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:09.625 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:09.625 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:09.625 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:09.625 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:09.625 "name": "raid_bdev1", 00:30:09.625 "uuid": "3524d2e1-6a61-4f7c-ad81-b96908001c6f", 00:30:09.625 "strip_size_kb": 0, 00:30:09.625 "state": "online", 00:30:09.625 "raid_level": "raid1", 00:30:09.625 "superblock": true, 00:30:09.625 "num_base_bdevs": 4, 00:30:09.625 "num_base_bdevs_discovered": 3, 00:30:09.625 "num_base_bdevs_operational": 3, 00:30:09.625 "base_bdevs_list": [ 00:30:09.625 { 00:30:09.625 "name": null, 00:30:09.625 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:09.625 "is_configured": false, 00:30:09.625 "data_offset": 0, 00:30:09.625 "data_size": 63488 00:30:09.625 }, 00:30:09.625 { 00:30:09.625 "name": "pt2", 00:30:09.625 "uuid": "00000000-0000-0000-0000-000000000002", 00:30:09.625 "is_configured": true, 00:30:09.625 "data_offset": 2048, 00:30:09.625 "data_size": 63488 00:30:09.625 }, 00:30:09.625 { 00:30:09.625 "name": "pt3", 00:30:09.625 "uuid": "00000000-0000-0000-0000-000000000003", 00:30:09.625 "is_configured": true, 00:30:09.625 "data_offset": 2048, 00:30:09.625 "data_size": 63488 00:30:09.625 }, 00:30:09.625 { 00:30:09.625 "name": "pt4", 00:30:09.625 "uuid": "00000000-0000-0000-0000-000000000004", 00:30:09.625 "is_configured": true, 00:30:09.625 "data_offset": 2048, 00:30:09.625 "data_size": 63488 00:30:09.625 } 00:30:09.625 ] 00:30:09.625 }' 00:30:09.625 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:09.625 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:09.883 [2024-11-20 13:50:16.817727] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:30:09.883 [2024-11-20 13:50:16.817769] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:30:09.883 [2024-11-20 13:50:16.817848] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:30:09.883 [2024-11-20 13:50:16.817935] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:30:09.883 [2024-11-20 13:50:16.817945] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt4 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:09.883 [2024-11-20 13:50:16.885718] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:30:09.883 [2024-11-20 13:50:16.885780] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:09.883 [2024-11-20 13:50:16.885799] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:30:09.883 [2024-11-20 13:50:16.885809] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:09.883 [2024-11-20 13:50:16.888191] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:09.883 [2024-11-20 13:50:16.888405] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:30:09.883 [2024-11-20 13:50:16.888512] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:30:09.883 [2024-11-20 13:50:16.888561] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:30:09.883 pt2 00:30:09.883 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:09.884 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:30:09.884 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:09.884 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:09.884 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:09.884 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:09.884 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:30:09.884 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:09.884 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:09.884 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:09.884 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:09.884 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:09.884 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:09.884 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:09.884 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:09.884 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:09.884 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:09.884 "name": "raid_bdev1", 00:30:09.884 "uuid": "3524d2e1-6a61-4f7c-ad81-b96908001c6f", 00:30:09.884 "strip_size_kb": 0, 00:30:09.884 "state": "configuring", 00:30:09.884 "raid_level": "raid1", 00:30:09.884 "superblock": true, 00:30:09.884 "num_base_bdevs": 4, 00:30:09.884 "num_base_bdevs_discovered": 1, 00:30:09.884 "num_base_bdevs_operational": 3, 00:30:09.884 "base_bdevs_list": [ 00:30:09.884 { 00:30:09.884 "name": null, 00:30:09.884 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:09.884 "is_configured": false, 00:30:09.884 "data_offset": 2048, 00:30:09.884 "data_size": 63488 00:30:09.884 }, 00:30:09.884 { 00:30:09.884 "name": "pt2", 00:30:09.884 "uuid": "00000000-0000-0000-0000-000000000002", 00:30:09.884 "is_configured": true, 00:30:09.884 "data_offset": 2048, 00:30:09.884 "data_size": 63488 00:30:09.884 }, 00:30:09.884 { 00:30:09.884 "name": null, 00:30:09.884 "uuid": "00000000-0000-0000-0000-000000000003", 00:30:09.884 "is_configured": false, 00:30:09.884 "data_offset": 2048, 00:30:09.884 "data_size": 63488 00:30:09.884 }, 00:30:09.884 { 00:30:09.884 "name": null, 00:30:09.884 "uuid": "00000000-0000-0000-0000-000000000004", 00:30:09.884 "is_configured": false, 00:30:09.884 "data_offset": 2048, 00:30:09.884 "data_size": 63488 00:30:09.884 } 00:30:09.884 ] 00:30:09.884 }' 00:30:09.884 13:50:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:09.884 13:50:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:10.449 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:30:10.449 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:30:10.449 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:30:10.449 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:10.449 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:10.449 [2024-11-20 13:50:17.245847] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:30:10.449 [2024-11-20 13:50:17.245926] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:10.449 [2024-11-20 13:50:17.245947] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:30:10.449 [2024-11-20 13:50:17.245956] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:10.449 [2024-11-20 13:50:17.246433] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:10.449 [2024-11-20 13:50:17.246450] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:30:10.449 [2024-11-20 13:50:17.246534] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:30:10.449 [2024-11-20 13:50:17.246556] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:30:10.449 pt3 00:30:10.449 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:10.449 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:30:10.449 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:10.449 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:10.449 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:10.449 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:10.449 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:30:10.449 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:10.449 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:10.449 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:10.449 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:10.449 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:10.449 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:10.449 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:10.449 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:10.449 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:10.449 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:10.449 "name": "raid_bdev1", 00:30:10.449 "uuid": "3524d2e1-6a61-4f7c-ad81-b96908001c6f", 00:30:10.449 "strip_size_kb": 0, 00:30:10.449 "state": "configuring", 00:30:10.449 "raid_level": "raid1", 00:30:10.449 "superblock": true, 00:30:10.449 "num_base_bdevs": 4, 00:30:10.449 "num_base_bdevs_discovered": 2, 00:30:10.449 "num_base_bdevs_operational": 3, 00:30:10.449 "base_bdevs_list": [ 00:30:10.449 { 00:30:10.449 "name": null, 00:30:10.449 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:10.449 "is_configured": false, 00:30:10.449 "data_offset": 2048, 00:30:10.449 "data_size": 63488 00:30:10.449 }, 00:30:10.449 { 00:30:10.449 "name": "pt2", 00:30:10.449 "uuid": "00000000-0000-0000-0000-000000000002", 00:30:10.449 "is_configured": true, 00:30:10.449 "data_offset": 2048, 00:30:10.449 "data_size": 63488 00:30:10.449 }, 00:30:10.449 { 00:30:10.449 "name": "pt3", 00:30:10.449 "uuid": "00000000-0000-0000-0000-000000000003", 00:30:10.449 "is_configured": true, 00:30:10.449 "data_offset": 2048, 00:30:10.449 "data_size": 63488 00:30:10.449 }, 00:30:10.449 { 00:30:10.449 "name": null, 00:30:10.449 "uuid": "00000000-0000-0000-0000-000000000004", 00:30:10.449 "is_configured": false, 00:30:10.449 "data_offset": 2048, 00:30:10.449 "data_size": 63488 00:30:10.449 } 00:30:10.449 ] 00:30:10.449 }' 00:30:10.449 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:10.449 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:10.707 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:30:10.707 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:30:10.707 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=3 00:30:10.707 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:30:10.707 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:10.707 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:10.707 [2024-11-20 13:50:17.573943] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:30:10.707 [2024-11-20 13:50:17.574043] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:10.707 [2024-11-20 13:50:17.574069] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:30:10.707 [2024-11-20 13:50:17.574079] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:10.707 [2024-11-20 13:50:17.574541] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:10.707 [2024-11-20 13:50:17.574568] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:30:10.707 [2024-11-20 13:50:17.574654] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:30:10.707 [2024-11-20 13:50:17.574676] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:30:10.707 [2024-11-20 13:50:17.574811] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:30:10.707 [2024-11-20 13:50:17.574825] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:30:10.707 [2024-11-20 13:50:17.575091] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:30:10.707 [2024-11-20 13:50:17.575415] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:30:10.707 [2024-11-20 13:50:17.575432] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:30:10.707 [2024-11-20 13:50:17.575575] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:10.707 pt4 00:30:10.707 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:10.707 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:30:10.708 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:10.708 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:10.708 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:10.708 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:10.708 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:30:10.708 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:10.708 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:10.708 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:10.708 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:10.708 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:10.708 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:10.708 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:10.708 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:10.708 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:10.708 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:10.708 "name": "raid_bdev1", 00:30:10.708 "uuid": "3524d2e1-6a61-4f7c-ad81-b96908001c6f", 00:30:10.708 "strip_size_kb": 0, 00:30:10.708 "state": "online", 00:30:10.708 "raid_level": "raid1", 00:30:10.708 "superblock": true, 00:30:10.708 "num_base_bdevs": 4, 00:30:10.708 "num_base_bdevs_discovered": 3, 00:30:10.708 "num_base_bdevs_operational": 3, 00:30:10.708 "base_bdevs_list": [ 00:30:10.708 { 00:30:10.708 "name": null, 00:30:10.708 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:10.708 "is_configured": false, 00:30:10.708 "data_offset": 2048, 00:30:10.708 "data_size": 63488 00:30:10.708 }, 00:30:10.708 { 00:30:10.708 "name": "pt2", 00:30:10.708 "uuid": "00000000-0000-0000-0000-000000000002", 00:30:10.708 "is_configured": true, 00:30:10.708 "data_offset": 2048, 00:30:10.708 "data_size": 63488 00:30:10.708 }, 00:30:10.708 { 00:30:10.708 "name": "pt3", 00:30:10.708 "uuid": "00000000-0000-0000-0000-000000000003", 00:30:10.708 "is_configured": true, 00:30:10.708 "data_offset": 2048, 00:30:10.708 "data_size": 63488 00:30:10.708 }, 00:30:10.708 { 00:30:10.708 "name": "pt4", 00:30:10.708 "uuid": "00000000-0000-0000-0000-000000000004", 00:30:10.708 "is_configured": true, 00:30:10.708 "data_offset": 2048, 00:30:10.708 "data_size": 63488 00:30:10.708 } 00:30:10.708 ] 00:30:10.708 }' 00:30:10.708 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:10.708 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:10.966 [2024-11-20 13:50:17.917968] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:30:10.966 [2024-11-20 13:50:17.918217] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:30:10.966 [2024-11-20 13:50:17.918315] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:30:10.966 [2024-11-20 13:50:17.918403] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:30:10.966 [2024-11-20 13:50:17.918416] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 4 -gt 2 ']' 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@534 -- # i=3 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt4 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:10.966 [2024-11-20 13:50:17.961971] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:30:10.966 [2024-11-20 13:50:17.962065] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:10.966 [2024-11-20 13:50:17.962082] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c080 00:30:10.966 [2024-11-20 13:50:17.962098] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:10.966 [2024-11-20 13:50:17.964580] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:10.966 [2024-11-20 13:50:17.964622] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:30:10.966 [2024-11-20 13:50:17.964716] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:30:10.966 [2024-11-20 13:50:17.964765] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:30:10.966 [2024-11-20 13:50:17.964902] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:30:10.966 [2024-11-20 13:50:17.964915] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:30:10.966 [2024-11-20 13:50:17.964932] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:30:10.966 [2024-11-20 13:50:17.965021] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:30:10.966 [2024-11-20 13:50:17.965130] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:30:10.966 pt1 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 4 -gt 2 ']' 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:10.966 "name": "raid_bdev1", 00:30:10.966 "uuid": "3524d2e1-6a61-4f7c-ad81-b96908001c6f", 00:30:10.966 "strip_size_kb": 0, 00:30:10.966 "state": "configuring", 00:30:10.966 "raid_level": "raid1", 00:30:10.966 "superblock": true, 00:30:10.966 "num_base_bdevs": 4, 00:30:10.966 "num_base_bdevs_discovered": 2, 00:30:10.966 "num_base_bdevs_operational": 3, 00:30:10.966 "base_bdevs_list": [ 00:30:10.966 { 00:30:10.966 "name": null, 00:30:10.966 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:10.966 "is_configured": false, 00:30:10.966 "data_offset": 2048, 00:30:10.966 "data_size": 63488 00:30:10.966 }, 00:30:10.966 { 00:30:10.966 "name": "pt2", 00:30:10.966 "uuid": "00000000-0000-0000-0000-000000000002", 00:30:10.966 "is_configured": true, 00:30:10.966 "data_offset": 2048, 00:30:10.966 "data_size": 63488 00:30:10.966 }, 00:30:10.966 { 00:30:10.966 "name": "pt3", 00:30:10.966 "uuid": "00000000-0000-0000-0000-000000000003", 00:30:10.966 "is_configured": true, 00:30:10.966 "data_offset": 2048, 00:30:10.966 "data_size": 63488 00:30:10.966 }, 00:30:10.966 { 00:30:10.966 "name": null, 00:30:10.966 "uuid": "00000000-0000-0000-0000-000000000004", 00:30:10.966 "is_configured": false, 00:30:10.966 "data_offset": 2048, 00:30:10.966 "data_size": 63488 00:30:10.966 } 00:30:10.966 ] 00:30:10.966 }' 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:10.966 13:50:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:11.532 13:50:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:30:11.532 13:50:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:30:11.532 13:50:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:11.532 13:50:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:11.532 13:50:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:11.532 13:50:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:30:11.532 13:50:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:30:11.532 13:50:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:11.532 13:50:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:11.532 [2024-11-20 13:50:18.322057] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:30:11.532 [2024-11-20 13:50:18.322130] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:11.532 [2024-11-20 13:50:18.322150] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c680 00:30:11.532 [2024-11-20 13:50:18.322159] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:11.532 [2024-11-20 13:50:18.322558] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:11.532 [2024-11-20 13:50:18.322576] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:30:11.532 [2024-11-20 13:50:18.322652] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:30:11.532 [2024-11-20 13:50:18.322672] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:30:11.532 [2024-11-20 13:50:18.322784] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:30:11.532 [2024-11-20 13:50:18.322791] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:30:11.532 [2024-11-20 13:50:18.323023] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:30:11.532 [2024-11-20 13:50:18.323146] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:30:11.532 [2024-11-20 13:50:18.323161] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:30:11.532 [2024-11-20 13:50:18.323276] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:11.532 pt4 00:30:11.532 13:50:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:11.533 13:50:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:30:11.533 13:50:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:11.533 13:50:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:11.533 13:50:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:11.533 13:50:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:11.533 13:50:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:30:11.533 13:50:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:11.533 13:50:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:11.533 13:50:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:11.533 13:50:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:11.533 13:50:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:11.533 13:50:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:11.533 13:50:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:11.533 13:50:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:11.533 13:50:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:11.533 13:50:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:11.533 "name": "raid_bdev1", 00:30:11.533 "uuid": "3524d2e1-6a61-4f7c-ad81-b96908001c6f", 00:30:11.533 "strip_size_kb": 0, 00:30:11.533 "state": "online", 00:30:11.533 "raid_level": "raid1", 00:30:11.533 "superblock": true, 00:30:11.533 "num_base_bdevs": 4, 00:30:11.533 "num_base_bdevs_discovered": 3, 00:30:11.533 "num_base_bdevs_operational": 3, 00:30:11.533 "base_bdevs_list": [ 00:30:11.533 { 00:30:11.533 "name": null, 00:30:11.533 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:11.533 "is_configured": false, 00:30:11.533 "data_offset": 2048, 00:30:11.533 "data_size": 63488 00:30:11.533 }, 00:30:11.533 { 00:30:11.533 "name": "pt2", 00:30:11.533 "uuid": "00000000-0000-0000-0000-000000000002", 00:30:11.533 "is_configured": true, 00:30:11.533 "data_offset": 2048, 00:30:11.533 "data_size": 63488 00:30:11.533 }, 00:30:11.533 { 00:30:11.533 "name": "pt3", 00:30:11.533 "uuid": "00000000-0000-0000-0000-000000000003", 00:30:11.533 "is_configured": true, 00:30:11.533 "data_offset": 2048, 00:30:11.533 "data_size": 63488 00:30:11.533 }, 00:30:11.533 { 00:30:11.533 "name": "pt4", 00:30:11.533 "uuid": "00000000-0000-0000-0000-000000000004", 00:30:11.533 "is_configured": true, 00:30:11.533 "data_offset": 2048, 00:30:11.533 "data_size": 63488 00:30:11.533 } 00:30:11.533 ] 00:30:11.533 }' 00:30:11.533 13:50:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:11.533 13:50:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:11.791 13:50:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:30:11.791 13:50:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:30:11.791 13:50:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:11.791 13:50:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:11.791 13:50:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:11.791 13:50:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:30:11.791 13:50:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:30:11.791 13:50:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:11.791 13:50:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:30:11.791 13:50:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:11.791 [2024-11-20 13:50:18.686403] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:30:11.791 13:50:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:11.791 13:50:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 3524d2e1-6a61-4f7c-ad81-b96908001c6f '!=' 3524d2e1-6a61-4f7c-ad81-b96908001c6f ']' 00:30:11.791 13:50:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 72470 00:30:11.791 13:50:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 72470 ']' 00:30:11.791 13:50:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 72470 00:30:11.791 13:50:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:30:11.791 13:50:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:30:11.791 13:50:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72470 00:30:11.791 13:50:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:30:11.791 13:50:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:30:11.791 13:50:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72470' 00:30:11.791 killing process with pid 72470 00:30:11.791 13:50:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 72470 00:30:11.791 [2024-11-20 13:50:18.729113] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:30:11.791 13:50:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 72470 00:30:11.791 [2024-11-20 13:50:18.729216] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:30:11.791 [2024-11-20 13:50:18.729290] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:30:11.791 [2024-11-20 13:50:18.729307] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:30:12.049 [2024-11-20 13:50:18.936003] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:30:12.618 13:50:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:30:12.618 ************************************ 00:30:12.618 END TEST raid_superblock_test 00:30:12.618 ************************************ 00:30:12.618 00:30:12.618 real 0m6.178s 00:30:12.618 user 0m9.781s 00:30:12.618 sys 0m1.088s 00:30:12.618 13:50:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:30:12.618 13:50:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:12.618 13:50:19 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 4 read 00:30:12.618 13:50:19 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:30:12.618 13:50:19 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:30:12.618 13:50:19 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:30:12.618 ************************************ 00:30:12.618 START TEST raid_read_error_test 00:30:12.618 ************************************ 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 4 read 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:30:12.618 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.LPhKbRL5yO 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=72932 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 72932 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 72932 ']' 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:12.618 13:50:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:30:12.618 [2024-11-20 13:50:19.654643] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:30:12.618 [2024-11-20 13:50:19.654760] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72932 ] 00:30:12.876 [2024-11-20 13:50:19.806850] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:30:12.876 [2024-11-20 13:50:19.914607] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:30:13.135 [2024-11-20 13:50:20.042201] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:30:13.135 [2024-11-20 13:50:20.042268] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:13.701 BaseBdev1_malloc 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:13.701 true 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:13.701 [2024-11-20 13:50:20.598525] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:30:13.701 [2024-11-20 13:50:20.598590] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:13.701 [2024-11-20 13:50:20.598608] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:30:13.701 [2024-11-20 13:50:20.598618] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:13.701 [2024-11-20 13:50:20.600567] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:13.701 [2024-11-20 13:50:20.600600] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:30:13.701 BaseBdev1 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:13.701 BaseBdev2_malloc 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:13.701 true 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:13.701 [2024-11-20 13:50:20.645281] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:30:13.701 [2024-11-20 13:50:20.645494] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:13.701 [2024-11-20 13:50:20.645517] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:30:13.701 [2024-11-20 13:50:20.645527] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:13.701 [2024-11-20 13:50:20.647497] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:13.701 [2024-11-20 13:50:20.647534] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:30:13.701 BaseBdev2 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:13.701 BaseBdev3_malloc 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:13.701 true 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:13.701 [2024-11-20 13:50:20.706000] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:30:13.701 [2024-11-20 13:50:20.706055] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:13.701 [2024-11-20 13:50:20.706072] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:30:13.701 [2024-11-20 13:50:20.706082] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:13.701 [2024-11-20 13:50:20.708045] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:13.701 [2024-11-20 13:50:20.708078] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:30:13.701 BaseBdev3 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:13.701 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:30:13.702 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:30:13.702 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:13.702 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:13.702 BaseBdev4_malloc 00:30:13.702 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:13.702 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:30:13.702 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:13.702 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:13.702 true 00:30:13.702 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:13.702 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:30:13.702 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:13.702 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:13.702 [2024-11-20 13:50:20.748411] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:30:13.702 [2024-11-20 13:50:20.748470] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:13.702 [2024-11-20 13:50:20.748486] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:30:13.702 [2024-11-20 13:50:20.748495] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:13.702 [2024-11-20 13:50:20.750418] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:13.702 [2024-11-20 13:50:20.750451] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:30:13.702 BaseBdev4 00:30:13.702 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:13.702 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:30:13.702 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:13.702 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:13.702 [2024-11-20 13:50:20.756481] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:30:13.960 [2024-11-20 13:50:20.758155] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:30:13.960 [2024-11-20 13:50:20.758222] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:30:13.960 [2024-11-20 13:50:20.758279] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:30:13.960 [2024-11-20 13:50:20.758477] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008580 00:30:13.960 [2024-11-20 13:50:20.758488] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:30:13.960 [2024-11-20 13:50:20.758700] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000068a0 00:30:13.960 [2024-11-20 13:50:20.758836] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008580 00:30:13.960 [2024-11-20 13:50:20.758844] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008580 00:30:13.960 [2024-11-20 13:50:20.758960] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:13.960 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:13.960 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:30:13.960 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:13.960 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:13.960 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:13.960 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:13.960 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:13.960 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:13.960 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:13.960 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:13.960 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:13.960 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:13.960 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:13.960 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:13.960 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:13.960 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:13.960 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:13.960 "name": "raid_bdev1", 00:30:13.960 "uuid": "e65d4e12-5a4b-4f5f-9288-34b0ff9249dd", 00:30:13.960 "strip_size_kb": 0, 00:30:13.960 "state": "online", 00:30:13.960 "raid_level": "raid1", 00:30:13.960 "superblock": true, 00:30:13.960 "num_base_bdevs": 4, 00:30:13.960 "num_base_bdevs_discovered": 4, 00:30:13.960 "num_base_bdevs_operational": 4, 00:30:13.960 "base_bdevs_list": [ 00:30:13.960 { 00:30:13.960 "name": "BaseBdev1", 00:30:13.960 "uuid": "665c24cc-4f9e-5fb8-8e2e-583dd624bcda", 00:30:13.960 "is_configured": true, 00:30:13.960 "data_offset": 2048, 00:30:13.960 "data_size": 63488 00:30:13.960 }, 00:30:13.960 { 00:30:13.960 "name": "BaseBdev2", 00:30:13.960 "uuid": "4bb098e5-98df-5ff3-a982-f9e7f2b9aa62", 00:30:13.960 "is_configured": true, 00:30:13.960 "data_offset": 2048, 00:30:13.960 "data_size": 63488 00:30:13.960 }, 00:30:13.960 { 00:30:13.960 "name": "BaseBdev3", 00:30:13.960 "uuid": "331bd803-5c5f-5359-8a93-c95001cadc17", 00:30:13.960 "is_configured": true, 00:30:13.960 "data_offset": 2048, 00:30:13.960 "data_size": 63488 00:30:13.960 }, 00:30:13.960 { 00:30:13.960 "name": "BaseBdev4", 00:30:13.960 "uuid": "8c318147-b450-51ae-bded-e644bfbf77c6", 00:30:13.960 "is_configured": true, 00:30:13.960 "data_offset": 2048, 00:30:13.960 "data_size": 63488 00:30:13.960 } 00:30:13.960 ] 00:30:13.960 }' 00:30:13.960 13:50:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:13.960 13:50:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:14.217 13:50:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:30:14.217 13:50:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:30:14.217 [2024-11-20 13:50:21.197469] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006a40 00:30:15.151 13:50:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:30:15.151 13:50:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:15.151 13:50:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:15.151 13:50:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:15.151 13:50:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:30:15.151 13:50:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:30:15.151 13:50:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:30:15.151 13:50:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:30:15.151 13:50:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:30:15.152 13:50:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:15.152 13:50:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:15.152 13:50:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:15.152 13:50:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:15.152 13:50:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:15.152 13:50:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:15.152 13:50:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:15.152 13:50:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:15.152 13:50:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:15.152 13:50:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:15.152 13:50:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:15.152 13:50:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:15.152 13:50:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:15.152 13:50:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:15.152 13:50:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:15.152 "name": "raid_bdev1", 00:30:15.152 "uuid": "e65d4e12-5a4b-4f5f-9288-34b0ff9249dd", 00:30:15.152 "strip_size_kb": 0, 00:30:15.152 "state": "online", 00:30:15.152 "raid_level": "raid1", 00:30:15.152 "superblock": true, 00:30:15.152 "num_base_bdevs": 4, 00:30:15.152 "num_base_bdevs_discovered": 4, 00:30:15.152 "num_base_bdevs_operational": 4, 00:30:15.152 "base_bdevs_list": [ 00:30:15.152 { 00:30:15.152 "name": "BaseBdev1", 00:30:15.152 "uuid": "665c24cc-4f9e-5fb8-8e2e-583dd624bcda", 00:30:15.152 "is_configured": true, 00:30:15.152 "data_offset": 2048, 00:30:15.152 "data_size": 63488 00:30:15.152 }, 00:30:15.152 { 00:30:15.152 "name": "BaseBdev2", 00:30:15.152 "uuid": "4bb098e5-98df-5ff3-a982-f9e7f2b9aa62", 00:30:15.152 "is_configured": true, 00:30:15.152 "data_offset": 2048, 00:30:15.152 "data_size": 63488 00:30:15.152 }, 00:30:15.152 { 00:30:15.152 "name": "BaseBdev3", 00:30:15.152 "uuid": "331bd803-5c5f-5359-8a93-c95001cadc17", 00:30:15.152 "is_configured": true, 00:30:15.152 "data_offset": 2048, 00:30:15.152 "data_size": 63488 00:30:15.152 }, 00:30:15.152 { 00:30:15.152 "name": "BaseBdev4", 00:30:15.152 "uuid": "8c318147-b450-51ae-bded-e644bfbf77c6", 00:30:15.152 "is_configured": true, 00:30:15.152 "data_offset": 2048, 00:30:15.152 "data_size": 63488 00:30:15.152 } 00:30:15.152 ] 00:30:15.152 }' 00:30:15.152 13:50:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:15.152 13:50:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:15.410 13:50:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:30:15.410 13:50:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:15.410 13:50:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:15.410 [2024-11-20 13:50:22.424784] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:30:15.410 [2024-11-20 13:50:22.424825] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:30:15.410 [2024-11-20 13:50:22.427382] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:30:15.410 [2024-11-20 13:50:22.427597] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:15.410 [2024-11-20 13:50:22.427724] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:30:15.410 [2024-11-20 13:50:22.427735] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state offline 00:30:15.410 { 00:30:15.410 "results": [ 00:30:15.410 { 00:30:15.410 "job": "raid_bdev1", 00:30:15.410 "core_mask": "0x1", 00:30:15.410 "workload": "randrw", 00:30:15.410 "percentage": 50, 00:30:15.410 "status": "finished", 00:30:15.410 "queue_depth": 1, 00:30:15.410 "io_size": 131072, 00:30:15.410 "runtime": 1.225698, 00:30:15.410 "iops": 11342.924602960926, 00:30:15.410 "mibps": 1417.8655753701157, 00:30:15.410 "io_failed": 0, 00:30:15.410 "io_timeout": 0, 00:30:15.410 "avg_latency_us": 85.54905095192515, 00:30:15.410 "min_latency_us": 23.926153846153845, 00:30:15.410 "max_latency_us": 1424.1476923076923 00:30:15.410 } 00:30:15.410 ], 00:30:15.410 "core_count": 1 00:30:15.410 } 00:30:15.410 13:50:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:15.410 13:50:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 72932 00:30:15.410 13:50:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 72932 ']' 00:30:15.410 13:50:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 72932 00:30:15.410 13:50:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:30:15.410 13:50:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:30:15.410 13:50:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72932 00:30:15.410 killing process with pid 72932 00:30:15.410 13:50:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:30:15.410 13:50:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:30:15.410 13:50:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72932' 00:30:15.410 13:50:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 72932 00:30:15.410 13:50:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 72932 00:30:15.410 [2024-11-20 13:50:22.452831] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:30:15.669 [2024-11-20 13:50:22.622927] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:30:16.237 13:50:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.LPhKbRL5yO 00:30:16.237 13:50:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:30:16.237 13:50:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:30:16.237 13:50:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:30:16.237 13:50:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:30:16.237 13:50:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:30:16.237 13:50:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:30:16.237 ************************************ 00:30:16.237 END TEST raid_read_error_test 00:30:16.237 ************************************ 00:30:16.237 13:50:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:30:16.237 00:30:16.237 real 0m3.700s 00:30:16.237 user 0m4.413s 00:30:16.237 sys 0m0.465s 00:30:16.237 13:50:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:30:16.237 13:50:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:16.495 13:50:23 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 4 write 00:30:16.495 13:50:23 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:30:16.495 13:50:23 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:30:16.495 13:50:23 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:30:16.495 ************************************ 00:30:16.495 START TEST raid_write_error_test 00:30:16.495 ************************************ 00:30:16.495 13:50:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 4 write 00:30:16.495 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:30:16.495 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:30:16.495 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:30:16.495 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:30:16.495 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:30:16.495 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:30:16.495 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:30:16.495 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:30:16.495 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:30:16.495 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:30:16.495 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:30:16.495 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:30:16.495 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:30:16.495 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:30:16.495 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:30:16.495 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:30:16.495 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:30:16.495 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:30:16.495 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:30:16.495 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:30:16.495 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:30:16.495 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:30:16.495 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:30:16.496 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:30:16.496 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:30:16.496 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:30:16.496 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:30:16.496 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.fiHoI04lxa 00:30:16.496 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=73072 00:30:16.496 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 73072 00:30:16.496 13:50:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 73072 ']' 00:30:16.496 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:30:16.496 13:50:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:30:16.496 13:50:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:30:16.496 13:50:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:30:16.496 13:50:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:30:16.496 13:50:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:16.496 13:50:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:30:16.496 [2024-11-20 13:50:23.407867] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:30:16.496 [2024-11-20 13:50:23.408011] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73072 ] 00:30:16.754 [2024-11-20 13:50:23.562577] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:30:16.754 [2024-11-20 13:50:23.680127] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:30:17.012 [2024-11-20 13:50:23.830833] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:30:17.012 [2024-11-20 13:50:23.830883] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:30:17.270 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:30:17.270 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:30:17.270 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:30:17.270 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:30:17.270 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:17.270 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:17.270 BaseBdev1_malloc 00:30:17.270 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:17.270 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:30:17.270 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:17.270 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:17.270 true 00:30:17.270 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:17.270 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:30:17.270 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:17.270 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:17.270 [2024-11-20 13:50:24.291865] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:30:17.270 [2024-11-20 13:50:24.291929] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:17.270 [2024-11-20 13:50:24.291950] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:30:17.270 [2024-11-20 13:50:24.291962] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:17.270 [2024-11-20 13:50:24.294248] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:17.270 [2024-11-20 13:50:24.294288] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:30:17.270 BaseBdev1 00:30:17.270 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:17.270 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:30:17.270 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:30:17.270 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:17.270 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:17.528 BaseBdev2_malloc 00:30:17.528 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:17.528 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:30:17.528 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:17.528 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:17.528 true 00:30:17.528 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:17.528 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:30:17.528 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:17.528 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:17.528 [2024-11-20 13:50:24.338527] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:30:17.528 [2024-11-20 13:50:24.338583] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:17.528 [2024-11-20 13:50:24.338598] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:30:17.528 [2024-11-20 13:50:24.338608] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:17.528 [2024-11-20 13:50:24.340928] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:17.528 [2024-11-20 13:50:24.341662] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:30:17.528 BaseBdev2 00:30:17.528 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:17.528 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:30:17.528 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:30:17.528 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:17.528 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:17.528 BaseBdev3_malloc 00:30:17.528 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:17.528 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:30:17.528 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:17.528 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:17.528 true 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:17.529 [2024-11-20 13:50:24.401835] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:30:17.529 [2024-11-20 13:50:24.402058] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:17.529 [2024-11-20 13:50:24.402084] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:30:17.529 [2024-11-20 13:50:24.402095] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:17.529 [2024-11-20 13:50:24.404389] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:17.529 [2024-11-20 13:50:24.404427] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:30:17.529 BaseBdev3 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:17.529 BaseBdev4_malloc 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:17.529 true 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:17.529 [2024-11-20 13:50:24.448469] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:30:17.529 [2024-11-20 13:50:24.448525] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:17.529 [2024-11-20 13:50:24.448544] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:30:17.529 [2024-11-20 13:50:24.448555] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:17.529 [2024-11-20 13:50:24.450818] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:17.529 [2024-11-20 13:50:24.450858] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:30:17.529 BaseBdev4 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:17.529 [2024-11-20 13:50:24.456546] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:30:17.529 [2024-11-20 13:50:24.458515] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:30:17.529 [2024-11-20 13:50:24.458590] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:30:17.529 [2024-11-20 13:50:24.458656] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:30:17.529 [2024-11-20 13:50:24.458887] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008580 00:30:17.529 [2024-11-20 13:50:24.458899] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:30:17.529 [2024-11-20 13:50:24.459161] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000068a0 00:30:17.529 [2024-11-20 13:50:24.459323] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008580 00:30:17.529 [2024-11-20 13:50:24.459339] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008580 00:30:17.529 [2024-11-20 13:50:24.459479] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:17.529 "name": "raid_bdev1", 00:30:17.529 "uuid": "681dbc10-6521-4483-a652-e54d935a1d7f", 00:30:17.529 "strip_size_kb": 0, 00:30:17.529 "state": "online", 00:30:17.529 "raid_level": "raid1", 00:30:17.529 "superblock": true, 00:30:17.529 "num_base_bdevs": 4, 00:30:17.529 "num_base_bdevs_discovered": 4, 00:30:17.529 "num_base_bdevs_operational": 4, 00:30:17.529 "base_bdevs_list": [ 00:30:17.529 { 00:30:17.529 "name": "BaseBdev1", 00:30:17.529 "uuid": "e4800bc5-c814-54df-a3c0-5c2492681855", 00:30:17.529 "is_configured": true, 00:30:17.529 "data_offset": 2048, 00:30:17.529 "data_size": 63488 00:30:17.529 }, 00:30:17.529 { 00:30:17.529 "name": "BaseBdev2", 00:30:17.529 "uuid": "ca3adb09-362f-52c6-b4db-c031915e561a", 00:30:17.529 "is_configured": true, 00:30:17.529 "data_offset": 2048, 00:30:17.529 "data_size": 63488 00:30:17.529 }, 00:30:17.529 { 00:30:17.529 "name": "BaseBdev3", 00:30:17.529 "uuid": "23bfe1ce-9679-5398-8720-e7ff78588916", 00:30:17.529 "is_configured": true, 00:30:17.529 "data_offset": 2048, 00:30:17.529 "data_size": 63488 00:30:17.529 }, 00:30:17.529 { 00:30:17.529 "name": "BaseBdev4", 00:30:17.529 "uuid": "b3d895d5-6f1f-5d60-b74e-1aa253c139b9", 00:30:17.529 "is_configured": true, 00:30:17.529 "data_offset": 2048, 00:30:17.529 "data_size": 63488 00:30:17.529 } 00:30:17.529 ] 00:30:17.529 }' 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:17.529 13:50:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:17.790 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:30:17.790 13:50:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:30:18.051 [2024-11-20 13:50:24.865714] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006a40 00:30:18.987 13:50:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:30:18.987 13:50:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:18.987 13:50:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:18.987 [2024-11-20 13:50:25.783994] bdev_raid.c:2276:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:30:18.987 [2024-11-20 13:50:25.784216] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:30:18.987 [2024-11-20 13:50:25.784484] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006a40 00:30:18.987 13:50:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:18.987 13:50:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:30:18.987 13:50:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:30:18.987 13:50:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:30:18.987 13:50:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=3 00:30:18.987 13:50:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:30:18.987 13:50:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:18.987 13:50:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:18.987 13:50:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:18.987 13:50:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:18.987 13:50:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:30:18.987 13:50:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:18.987 13:50:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:18.987 13:50:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:18.987 13:50:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:18.987 13:50:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:18.987 13:50:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:18.987 13:50:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:18.987 13:50:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:18.987 13:50:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:18.987 13:50:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:18.987 "name": "raid_bdev1", 00:30:18.987 "uuid": "681dbc10-6521-4483-a652-e54d935a1d7f", 00:30:18.987 "strip_size_kb": 0, 00:30:18.987 "state": "online", 00:30:18.987 "raid_level": "raid1", 00:30:18.987 "superblock": true, 00:30:18.987 "num_base_bdevs": 4, 00:30:18.987 "num_base_bdevs_discovered": 3, 00:30:18.987 "num_base_bdevs_operational": 3, 00:30:18.987 "base_bdevs_list": [ 00:30:18.987 { 00:30:18.987 "name": null, 00:30:18.987 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:18.987 "is_configured": false, 00:30:18.987 "data_offset": 0, 00:30:18.987 "data_size": 63488 00:30:18.987 }, 00:30:18.987 { 00:30:18.987 "name": "BaseBdev2", 00:30:18.987 "uuid": "ca3adb09-362f-52c6-b4db-c031915e561a", 00:30:18.987 "is_configured": true, 00:30:18.987 "data_offset": 2048, 00:30:18.987 "data_size": 63488 00:30:18.987 }, 00:30:18.987 { 00:30:18.987 "name": "BaseBdev3", 00:30:18.987 "uuid": "23bfe1ce-9679-5398-8720-e7ff78588916", 00:30:18.987 "is_configured": true, 00:30:18.987 "data_offset": 2048, 00:30:18.987 "data_size": 63488 00:30:18.988 }, 00:30:18.988 { 00:30:18.988 "name": "BaseBdev4", 00:30:18.988 "uuid": "b3d895d5-6f1f-5d60-b74e-1aa253c139b9", 00:30:18.988 "is_configured": true, 00:30:18.988 "data_offset": 2048, 00:30:18.988 "data_size": 63488 00:30:18.988 } 00:30:18.988 ] 00:30:18.988 }' 00:30:18.988 13:50:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:18.988 13:50:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:19.246 13:50:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:30:19.246 13:50:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:19.246 13:50:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:19.246 [2024-11-20 13:50:26.108259] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:30:19.246 [2024-11-20 13:50:26.108290] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:30:19.246 [2024-11-20 13:50:26.111447] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:30:19.246 [2024-11-20 13:50:26.111638] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:19.246 [2024-11-20 13:50:26.111769] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:30:19.246 [2024-11-20 13:50:26.111783] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state offline 00:30:19.246 { 00:30:19.246 "results": [ 00:30:19.246 { 00:30:19.246 "job": "raid_bdev1", 00:30:19.246 "core_mask": "0x1", 00:30:19.246 "workload": "randrw", 00:30:19.246 "percentage": 50, 00:30:19.246 "status": "finished", 00:30:19.246 "queue_depth": 1, 00:30:19.246 "io_size": 131072, 00:30:19.246 "runtime": 1.240588, 00:30:19.246 "iops": 10432.149916007571, 00:30:19.246 "mibps": 1304.0187395009464, 00:30:19.246 "io_failed": 0, 00:30:19.246 "io_timeout": 0, 00:30:19.246 "avg_latency_us": 92.56363753075854, 00:30:19.246 "min_latency_us": 30.326153846153847, 00:30:19.246 "max_latency_us": 1777.033846153846 00:30:19.246 } 00:30:19.246 ], 00:30:19.246 "core_count": 1 00:30:19.246 } 00:30:19.246 13:50:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:19.246 13:50:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 73072 00:30:19.246 13:50:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 73072 ']' 00:30:19.246 13:50:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 73072 00:30:19.246 13:50:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:30:19.246 13:50:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:30:19.246 13:50:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73072 00:30:19.246 killing process with pid 73072 00:30:19.246 13:50:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:30:19.246 13:50:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:30:19.246 13:50:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73072' 00:30:19.246 13:50:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 73072 00:30:19.246 [2024-11-20 13:50:26.138422] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:30:19.246 13:50:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 73072 00:30:19.504 [2024-11-20 13:50:26.310238] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:30:20.070 13:50:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:30:20.070 13:50:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.fiHoI04lxa 00:30:20.070 13:50:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:30:20.070 ************************************ 00:30:20.070 END TEST raid_write_error_test 00:30:20.070 ************************************ 00:30:20.070 13:50:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:30:20.070 13:50:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:30:20.070 13:50:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:30:20.070 13:50:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:30:20.070 13:50:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:30:20.070 00:30:20.070 real 0m3.656s 00:30:20.070 user 0m4.270s 00:30:20.070 sys 0m0.469s 00:30:20.070 13:50:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:30:20.070 13:50:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:30:20.070 13:50:27 bdev_raid -- bdev/bdev_raid.sh@976 -- # '[' true = true ']' 00:30:20.070 13:50:27 bdev_raid -- bdev/bdev_raid.sh@977 -- # for n in 2 4 00:30:20.070 13:50:27 bdev_raid -- bdev/bdev_raid.sh@978 -- # run_test raid_rebuild_test raid_rebuild_test raid1 2 false false true 00:30:20.070 13:50:27 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:30:20.070 13:50:27 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:30:20.070 13:50:27 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:30:20.070 ************************************ 00:30:20.070 START TEST raid_rebuild_test 00:30:20.070 ************************************ 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 false false true 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=73199 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 73199 00:30:20.070 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 73199 ']' 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:30:20.070 13:50:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:30:20.071 13:50:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:20.071 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:30:20.071 I/O size of 3145728 is greater than zero copy threshold (65536). 00:30:20.071 Zero copy mechanism will not be used. 00:30:20.071 [2024-11-20 13:50:27.112262] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:30:20.071 [2024-11-20 13:50:27.112392] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73199 ] 00:30:20.328 [2024-11-20 13:50:27.275399] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:30:20.584 [2024-11-20 13:50:27.396937] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:30:20.584 [2024-11-20 13:50:27.548167] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:30:20.584 [2024-11-20 13:50:27.548217] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:30:21.148 13:50:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:30:21.148 13:50:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:30:21.148 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:30:21.149 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:30:21.149 13:50:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:21.149 13:50:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:21.149 BaseBdev1_malloc 00:30:21.149 13:50:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:21.149 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:30:21.149 13:50:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:21.149 13:50:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:21.149 [2024-11-20 13:50:27.991576] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:30:21.149 [2024-11-20 13:50:27.991654] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:21.149 [2024-11-20 13:50:27.991679] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:30:21.149 [2024-11-20 13:50:27.991691] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:21.149 [2024-11-20 13:50:27.994033] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:21.149 [2024-11-20 13:50:27.994253] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:30:21.149 BaseBdev1 00:30:21.149 13:50:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:21.149 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:30:21.149 13:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:30:21.149 13:50:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:21.149 13:50:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:21.149 BaseBdev2_malloc 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:21.149 [2024-11-20 13:50:28.030669] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:30:21.149 [2024-11-20 13:50:28.030737] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:21.149 [2024-11-20 13:50:28.030759] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:30:21.149 [2024-11-20 13:50:28.030771] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:21.149 [2024-11-20 13:50:28.033057] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:21.149 [2024-11-20 13:50:28.033217] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:30:21.149 BaseBdev2 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:21.149 spare_malloc 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:21.149 spare_delay 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:21.149 [2024-11-20 13:50:28.087610] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:30:21.149 [2024-11-20 13:50:28.087677] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:21.149 [2024-11-20 13:50:28.087699] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:30:21.149 [2024-11-20 13:50:28.087712] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:21.149 [2024-11-20 13:50:28.090035] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:21.149 [2024-11-20 13:50:28.090071] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:30:21.149 spare 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:21.149 [2024-11-20 13:50:28.095662] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:30:21.149 [2024-11-20 13:50:28.097710] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:30:21.149 [2024-11-20 13:50:28.097961] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:30:21.149 [2024-11-20 13:50:28.097997] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:30:21.149 [2024-11-20 13:50:28.098269] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:30:21.149 [2024-11-20 13:50:28.098422] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:30:21.149 [2024-11-20 13:50:28.098432] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:30:21.149 [2024-11-20 13:50:28.098583] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:21.149 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:21.149 "name": "raid_bdev1", 00:30:21.149 "uuid": "bc6d1cf8-5f20-45e1-a77c-b6271acf8858", 00:30:21.149 "strip_size_kb": 0, 00:30:21.149 "state": "online", 00:30:21.149 "raid_level": "raid1", 00:30:21.149 "superblock": false, 00:30:21.149 "num_base_bdevs": 2, 00:30:21.149 "num_base_bdevs_discovered": 2, 00:30:21.149 "num_base_bdevs_operational": 2, 00:30:21.149 "base_bdevs_list": [ 00:30:21.149 { 00:30:21.149 "name": "BaseBdev1", 00:30:21.149 "uuid": "a0a1decd-df4f-5023-a35a-92a3808ccbc3", 00:30:21.149 "is_configured": true, 00:30:21.149 "data_offset": 0, 00:30:21.149 "data_size": 65536 00:30:21.149 }, 00:30:21.149 { 00:30:21.149 "name": "BaseBdev2", 00:30:21.149 "uuid": "ea7f79a4-e9a9-548b-8b55-aa9d2caee82d", 00:30:21.149 "is_configured": true, 00:30:21.149 "data_offset": 0, 00:30:21.149 "data_size": 65536 00:30:21.149 } 00:30:21.150 ] 00:30:21.150 }' 00:30:21.150 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:21.150 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:21.407 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:30:21.407 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:30:21.407 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:21.407 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:21.407 [2024-11-20 13:50:28.412121] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:30:21.407 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:21.407 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:30:21.407 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:30:21.407 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:21.407 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:21.407 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:21.407 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:21.664 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:30:21.664 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:30:21.664 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:30:21.665 [2024-11-20 13:50:28.663881] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:30:21.665 /dev/nbd0 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:30:21.665 1+0 records in 00:30:21.665 1+0 records out 00:30:21.665 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000384732 s, 10.6 MB/s 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:30:21.665 13:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=65536 oflag=direct 00:30:26.925 65536+0 records in 00:30:26.925 65536+0 records out 00:30:26.925 33554432 bytes (34 MB, 32 MiB) copied, 4.58265 s, 7.3 MB/s 00:30:26.925 13:50:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:30:26.925 13:50:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:30:26.925 13:50:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:30:26.925 13:50:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:30:26.925 13:50:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:30:26.925 13:50:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:30:26.926 [2024-11-20 13:50:33.494723] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:26.926 [2024-11-20 13:50:33.522803] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:26.926 "name": "raid_bdev1", 00:30:26.926 "uuid": "bc6d1cf8-5f20-45e1-a77c-b6271acf8858", 00:30:26.926 "strip_size_kb": 0, 00:30:26.926 "state": "online", 00:30:26.926 "raid_level": "raid1", 00:30:26.926 "superblock": false, 00:30:26.926 "num_base_bdevs": 2, 00:30:26.926 "num_base_bdevs_discovered": 1, 00:30:26.926 "num_base_bdevs_operational": 1, 00:30:26.926 "base_bdevs_list": [ 00:30:26.926 { 00:30:26.926 "name": null, 00:30:26.926 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:26.926 "is_configured": false, 00:30:26.926 "data_offset": 0, 00:30:26.926 "data_size": 65536 00:30:26.926 }, 00:30:26.926 { 00:30:26.926 "name": "BaseBdev2", 00:30:26.926 "uuid": "ea7f79a4-e9a9-548b-8b55-aa9d2caee82d", 00:30:26.926 "is_configured": true, 00:30:26.926 "data_offset": 0, 00:30:26.926 "data_size": 65536 00:30:26.926 } 00:30:26.926 ] 00:30:26.926 }' 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:26.926 [2024-11-20 13:50:33.834872] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:30:26.926 [2024-11-20 13:50:33.844995] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d09bd0 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:26.926 13:50:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:30:26.926 [2024-11-20 13:50:33.846695] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:30:27.858 13:50:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:30:27.858 13:50:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:27.858 13:50:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:30:27.858 13:50:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:30:27.858 13:50:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:27.858 13:50:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:27.858 13:50:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:27.858 13:50:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:27.858 13:50:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:27.858 13:50:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:27.858 13:50:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:27.858 "name": "raid_bdev1", 00:30:27.858 "uuid": "bc6d1cf8-5f20-45e1-a77c-b6271acf8858", 00:30:27.858 "strip_size_kb": 0, 00:30:27.858 "state": "online", 00:30:27.858 "raid_level": "raid1", 00:30:27.858 "superblock": false, 00:30:27.858 "num_base_bdevs": 2, 00:30:27.858 "num_base_bdevs_discovered": 2, 00:30:27.858 "num_base_bdevs_operational": 2, 00:30:27.858 "process": { 00:30:27.858 "type": "rebuild", 00:30:27.858 "target": "spare", 00:30:27.858 "progress": { 00:30:27.858 "blocks": 20480, 00:30:27.858 "percent": 31 00:30:27.858 } 00:30:27.858 }, 00:30:27.858 "base_bdevs_list": [ 00:30:27.858 { 00:30:27.858 "name": "spare", 00:30:27.858 "uuid": "16854160-4998-504d-9997-8d2a187ac681", 00:30:27.858 "is_configured": true, 00:30:27.858 "data_offset": 0, 00:30:27.858 "data_size": 65536 00:30:27.858 }, 00:30:27.858 { 00:30:27.858 "name": "BaseBdev2", 00:30:27.858 "uuid": "ea7f79a4-e9a9-548b-8b55-aa9d2caee82d", 00:30:27.858 "is_configured": true, 00:30:27.858 "data_offset": 0, 00:30:27.858 "data_size": 65536 00:30:27.858 } 00:30:27.858 ] 00:30:27.858 }' 00:30:27.858 13:50:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:27.858 13:50:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:30:28.116 13:50:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:28.116 13:50:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:30:28.116 13:50:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:30:28.116 13:50:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:28.116 13:50:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:28.116 [2024-11-20 13:50:34.940826] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:30:28.116 [2024-11-20 13:50:34.953359] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:30:28.116 [2024-11-20 13:50:34.953420] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:28.116 [2024-11-20 13:50:34.953433] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:30:28.116 [2024-11-20 13:50:34.953441] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:30:28.116 13:50:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:28.116 13:50:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:30:28.116 13:50:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:28.116 13:50:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:28.116 13:50:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:28.116 13:50:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:28.116 13:50:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:30:28.116 13:50:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:28.116 13:50:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:28.116 13:50:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:28.116 13:50:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:28.116 13:50:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:28.116 13:50:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:28.116 13:50:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:28.116 13:50:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:28.116 13:50:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:28.116 13:50:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:28.116 "name": "raid_bdev1", 00:30:28.116 "uuid": "bc6d1cf8-5f20-45e1-a77c-b6271acf8858", 00:30:28.116 "strip_size_kb": 0, 00:30:28.116 "state": "online", 00:30:28.116 "raid_level": "raid1", 00:30:28.116 "superblock": false, 00:30:28.116 "num_base_bdevs": 2, 00:30:28.116 "num_base_bdevs_discovered": 1, 00:30:28.116 "num_base_bdevs_operational": 1, 00:30:28.116 "base_bdevs_list": [ 00:30:28.116 { 00:30:28.116 "name": null, 00:30:28.116 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:28.116 "is_configured": false, 00:30:28.116 "data_offset": 0, 00:30:28.116 "data_size": 65536 00:30:28.116 }, 00:30:28.116 { 00:30:28.116 "name": "BaseBdev2", 00:30:28.116 "uuid": "ea7f79a4-e9a9-548b-8b55-aa9d2caee82d", 00:30:28.116 "is_configured": true, 00:30:28.116 "data_offset": 0, 00:30:28.116 "data_size": 65536 00:30:28.116 } 00:30:28.116 ] 00:30:28.116 }' 00:30:28.116 13:50:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:28.116 13:50:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:28.374 13:50:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:30:28.374 13:50:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:28.374 13:50:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:30:28.374 13:50:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:30:28.374 13:50:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:28.374 13:50:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:28.374 13:50:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:28.374 13:50:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:28.374 13:50:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:28.374 13:50:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:28.374 13:50:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:28.374 "name": "raid_bdev1", 00:30:28.374 "uuid": "bc6d1cf8-5f20-45e1-a77c-b6271acf8858", 00:30:28.374 "strip_size_kb": 0, 00:30:28.374 "state": "online", 00:30:28.374 "raid_level": "raid1", 00:30:28.374 "superblock": false, 00:30:28.374 "num_base_bdevs": 2, 00:30:28.374 "num_base_bdevs_discovered": 1, 00:30:28.374 "num_base_bdevs_operational": 1, 00:30:28.374 "base_bdevs_list": [ 00:30:28.374 { 00:30:28.374 "name": null, 00:30:28.374 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:28.374 "is_configured": false, 00:30:28.374 "data_offset": 0, 00:30:28.374 "data_size": 65536 00:30:28.374 }, 00:30:28.374 { 00:30:28.374 "name": "BaseBdev2", 00:30:28.374 "uuid": "ea7f79a4-e9a9-548b-8b55-aa9d2caee82d", 00:30:28.374 "is_configured": true, 00:30:28.374 "data_offset": 0, 00:30:28.374 "data_size": 65536 00:30:28.374 } 00:30:28.374 ] 00:30:28.374 }' 00:30:28.374 13:50:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:28.374 13:50:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:30:28.374 13:50:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:28.374 13:50:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:30:28.374 13:50:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:30:28.375 13:50:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:28.375 13:50:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:28.375 [2024-11-20 13:50:35.381384] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:30:28.375 [2024-11-20 13:50:35.390835] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d09ca0 00:30:28.375 13:50:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:28.375 13:50:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:30:28.375 [2024-11-20 13:50:35.392602] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:29.755 "name": "raid_bdev1", 00:30:29.755 "uuid": "bc6d1cf8-5f20-45e1-a77c-b6271acf8858", 00:30:29.755 "strip_size_kb": 0, 00:30:29.755 "state": "online", 00:30:29.755 "raid_level": "raid1", 00:30:29.755 "superblock": false, 00:30:29.755 "num_base_bdevs": 2, 00:30:29.755 "num_base_bdevs_discovered": 2, 00:30:29.755 "num_base_bdevs_operational": 2, 00:30:29.755 "process": { 00:30:29.755 "type": "rebuild", 00:30:29.755 "target": "spare", 00:30:29.755 "progress": { 00:30:29.755 "blocks": 18432, 00:30:29.755 "percent": 28 00:30:29.755 } 00:30:29.755 }, 00:30:29.755 "base_bdevs_list": [ 00:30:29.755 { 00:30:29.755 "name": "spare", 00:30:29.755 "uuid": "16854160-4998-504d-9997-8d2a187ac681", 00:30:29.755 "is_configured": true, 00:30:29.755 "data_offset": 0, 00:30:29.755 "data_size": 65536 00:30:29.755 }, 00:30:29.755 { 00:30:29.755 "name": "BaseBdev2", 00:30:29.755 "uuid": "ea7f79a4-e9a9-548b-8b55-aa9d2caee82d", 00:30:29.755 "is_configured": true, 00:30:29.755 "data_offset": 0, 00:30:29.755 "data_size": 65536 00:30:29.755 } 00:30:29.755 ] 00:30:29.755 }' 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=283 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:29.755 "name": "raid_bdev1", 00:30:29.755 "uuid": "bc6d1cf8-5f20-45e1-a77c-b6271acf8858", 00:30:29.755 "strip_size_kb": 0, 00:30:29.755 "state": "online", 00:30:29.755 "raid_level": "raid1", 00:30:29.755 "superblock": false, 00:30:29.755 "num_base_bdevs": 2, 00:30:29.755 "num_base_bdevs_discovered": 2, 00:30:29.755 "num_base_bdevs_operational": 2, 00:30:29.755 "process": { 00:30:29.755 "type": "rebuild", 00:30:29.755 "target": "spare", 00:30:29.755 "progress": { 00:30:29.755 "blocks": 22528, 00:30:29.755 "percent": 34 00:30:29.755 } 00:30:29.755 }, 00:30:29.755 "base_bdevs_list": [ 00:30:29.755 { 00:30:29.755 "name": "spare", 00:30:29.755 "uuid": "16854160-4998-504d-9997-8d2a187ac681", 00:30:29.755 "is_configured": true, 00:30:29.755 "data_offset": 0, 00:30:29.755 "data_size": 65536 00:30:29.755 }, 00:30:29.755 { 00:30:29.755 "name": "BaseBdev2", 00:30:29.755 "uuid": "ea7f79a4-e9a9-548b-8b55-aa9d2caee82d", 00:30:29.755 "is_configured": true, 00:30:29.755 "data_offset": 0, 00:30:29.755 "data_size": 65536 00:30:29.755 } 00:30:29.755 ] 00:30:29.755 }' 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:30:29.755 13:50:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:30:30.688 13:50:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:30:30.688 13:50:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:30:30.688 13:50:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:30.688 13:50:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:30:30.688 13:50:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:30:30.688 13:50:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:30.688 13:50:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:30.688 13:50:37 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:30.688 13:50:37 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:30.688 13:50:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:30.688 13:50:37 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:30.688 13:50:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:30.688 "name": "raid_bdev1", 00:30:30.688 "uuid": "bc6d1cf8-5f20-45e1-a77c-b6271acf8858", 00:30:30.688 "strip_size_kb": 0, 00:30:30.688 "state": "online", 00:30:30.688 "raid_level": "raid1", 00:30:30.688 "superblock": false, 00:30:30.688 "num_base_bdevs": 2, 00:30:30.688 "num_base_bdevs_discovered": 2, 00:30:30.688 "num_base_bdevs_operational": 2, 00:30:30.688 "process": { 00:30:30.688 "type": "rebuild", 00:30:30.688 "target": "spare", 00:30:30.688 "progress": { 00:30:30.688 "blocks": 43008, 00:30:30.688 "percent": 65 00:30:30.688 } 00:30:30.688 }, 00:30:30.688 "base_bdevs_list": [ 00:30:30.688 { 00:30:30.688 "name": "spare", 00:30:30.688 "uuid": "16854160-4998-504d-9997-8d2a187ac681", 00:30:30.688 "is_configured": true, 00:30:30.688 "data_offset": 0, 00:30:30.688 "data_size": 65536 00:30:30.688 }, 00:30:30.688 { 00:30:30.688 "name": "BaseBdev2", 00:30:30.688 "uuid": "ea7f79a4-e9a9-548b-8b55-aa9d2caee82d", 00:30:30.688 "is_configured": true, 00:30:30.688 "data_offset": 0, 00:30:30.688 "data_size": 65536 00:30:30.688 } 00:30:30.688 ] 00:30:30.688 }' 00:30:30.688 13:50:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:30.688 13:50:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:30:30.688 13:50:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:30.688 13:50:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:30:30.688 13:50:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:30:31.685 [2024-11-20 13:50:38.611765] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:30:31.685 [2024-11-20 13:50:38.612044] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:30:31.685 [2024-11-20 13:50:38.612098] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:31.685 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:30:31.685 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:30:31.685 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:31.685 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:30:31.685 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:30:31.685 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:31.685 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:31.685 13:50:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:31.685 13:50:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:31.685 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:31.685 13:50:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:31.944 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:31.944 "name": "raid_bdev1", 00:30:31.944 "uuid": "bc6d1cf8-5f20-45e1-a77c-b6271acf8858", 00:30:31.944 "strip_size_kb": 0, 00:30:31.944 "state": "online", 00:30:31.944 "raid_level": "raid1", 00:30:31.944 "superblock": false, 00:30:31.944 "num_base_bdevs": 2, 00:30:31.944 "num_base_bdevs_discovered": 2, 00:30:31.944 "num_base_bdevs_operational": 2, 00:30:31.944 "base_bdevs_list": [ 00:30:31.944 { 00:30:31.944 "name": "spare", 00:30:31.944 "uuid": "16854160-4998-504d-9997-8d2a187ac681", 00:30:31.944 "is_configured": true, 00:30:31.944 "data_offset": 0, 00:30:31.944 "data_size": 65536 00:30:31.944 }, 00:30:31.944 { 00:30:31.944 "name": "BaseBdev2", 00:30:31.944 "uuid": "ea7f79a4-e9a9-548b-8b55-aa9d2caee82d", 00:30:31.944 "is_configured": true, 00:30:31.944 "data_offset": 0, 00:30:31.944 "data_size": 65536 00:30:31.944 } 00:30:31.944 ] 00:30:31.944 }' 00:30:31.944 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:31.944 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:30:31.944 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:31.944 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:30:31.944 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:30:31.944 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:31.945 "name": "raid_bdev1", 00:30:31.945 "uuid": "bc6d1cf8-5f20-45e1-a77c-b6271acf8858", 00:30:31.945 "strip_size_kb": 0, 00:30:31.945 "state": "online", 00:30:31.945 "raid_level": "raid1", 00:30:31.945 "superblock": false, 00:30:31.945 "num_base_bdevs": 2, 00:30:31.945 "num_base_bdevs_discovered": 2, 00:30:31.945 "num_base_bdevs_operational": 2, 00:30:31.945 "base_bdevs_list": [ 00:30:31.945 { 00:30:31.945 "name": "spare", 00:30:31.945 "uuid": "16854160-4998-504d-9997-8d2a187ac681", 00:30:31.945 "is_configured": true, 00:30:31.945 "data_offset": 0, 00:30:31.945 "data_size": 65536 00:30:31.945 }, 00:30:31.945 { 00:30:31.945 "name": "BaseBdev2", 00:30:31.945 "uuid": "ea7f79a4-e9a9-548b-8b55-aa9d2caee82d", 00:30:31.945 "is_configured": true, 00:30:31.945 "data_offset": 0, 00:30:31.945 "data_size": 65536 00:30:31.945 } 00:30:31.945 ] 00:30:31.945 }' 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:31.945 "name": "raid_bdev1", 00:30:31.945 "uuid": "bc6d1cf8-5f20-45e1-a77c-b6271acf8858", 00:30:31.945 "strip_size_kb": 0, 00:30:31.945 "state": "online", 00:30:31.945 "raid_level": "raid1", 00:30:31.945 "superblock": false, 00:30:31.945 "num_base_bdevs": 2, 00:30:31.945 "num_base_bdevs_discovered": 2, 00:30:31.945 "num_base_bdevs_operational": 2, 00:30:31.945 "base_bdevs_list": [ 00:30:31.945 { 00:30:31.945 "name": "spare", 00:30:31.945 "uuid": "16854160-4998-504d-9997-8d2a187ac681", 00:30:31.945 "is_configured": true, 00:30:31.945 "data_offset": 0, 00:30:31.945 "data_size": 65536 00:30:31.945 }, 00:30:31.945 { 00:30:31.945 "name": "BaseBdev2", 00:30:31.945 "uuid": "ea7f79a4-e9a9-548b-8b55-aa9d2caee82d", 00:30:31.945 "is_configured": true, 00:30:31.945 "data_offset": 0, 00:30:31.945 "data_size": 65536 00:30:31.945 } 00:30:31.945 ] 00:30:31.945 }' 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:31.945 13:50:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:32.203 13:50:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:30:32.203 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:32.203 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:32.203 [2024-11-20 13:50:39.187364] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:30:32.203 [2024-11-20 13:50:39.187563] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:30:32.203 [2024-11-20 13:50:39.187657] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:30:32.203 [2024-11-20 13:50:39.187725] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:30:32.203 [2024-11-20 13:50:39.187735] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:30:32.203 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:32.203 13:50:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:30:32.203 13:50:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:32.203 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:32.203 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:32.203 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:32.203 13:50:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:30:32.203 13:50:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:30:32.203 13:50:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:30:32.203 13:50:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:30:32.203 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:30:32.203 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:30:32.203 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:30:32.203 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:30:32.203 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:30:32.203 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:30:32.203 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:30:32.203 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:30:32.203 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:30:32.461 /dev/nbd0 00:30:32.461 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:30:32.461 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:30:32.461 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:30:32.461 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:30:32.461 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:30:32.461 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:30:32.461 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:30:32.461 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:30:32.461 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:30:32.461 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:30:32.461 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:30:32.461 1+0 records in 00:30:32.461 1+0 records out 00:30:32.461 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000296818 s, 13.8 MB/s 00:30:32.461 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:30:32.461 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:30:32.462 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:30:32.462 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:30:32.462 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:30:32.462 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:30:32.462 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:30:32.462 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:30:32.719 /dev/nbd1 00:30:32.720 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:30:32.720 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:30:32.720 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:30:32.720 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:30:32.720 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:30:32.720 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:30:32.720 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:30:32.720 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:30:32.720 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:30:32.720 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:30:32.720 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:30:32.720 1+0 records in 00:30:32.720 1+0 records out 00:30:32.720 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000245215 s, 16.7 MB/s 00:30:32.720 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:30:32.720 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:30:32.720 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:30:32.720 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:30:32.720 13:50:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:30:32.720 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:30:32.720 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:30:32.720 13:50:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:30:32.977 13:50:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:30:32.978 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:30:32.978 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:30:32.978 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:30:32.978 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:30:32.978 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:30:32.978 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:30:32.978 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:30:32.978 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:30:32.978 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:30:32.978 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:30:32.978 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:30:32.978 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:30:32.978 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:30:32.978 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:30:32.978 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:30:32.978 13:50:39 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:30:33.235 13:50:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:30:33.235 13:50:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:30:33.235 13:50:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:30:33.235 13:50:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:30:33.235 13:50:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:30:33.235 13:50:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:30:33.235 13:50:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:30:33.235 13:50:40 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:30:33.235 13:50:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:30:33.235 13:50:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 73199 00:30:33.235 13:50:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 73199 ']' 00:30:33.235 13:50:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 73199 00:30:33.235 13:50:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:30:33.235 13:50:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:30:33.235 13:50:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73199 00:30:33.235 13:50:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:30:33.235 13:50:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:30:33.235 killing process with pid 73199 00:30:33.235 13:50:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73199' 00:30:33.235 13:50:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@973 -- # kill 73199 00:30:33.235 Received shutdown signal, test time was about 60.000000 seconds 00:30:33.235 00:30:33.235 Latency(us) 00:30:33.235 [2024-11-20T13:50:40.294Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:30:33.235 [2024-11-20T13:50:40.294Z] =================================================================================================================== 00:30:33.235 [2024-11-20T13:50:40.294Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:30:33.235 13:50:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@978 -- # wait 73199 00:30:33.235 [2024-11-20 13:50:40.172843] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:30:33.493 [2024-11-20 13:50:40.329618] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:30:34.139 ************************************ 00:30:34.139 END TEST raid_rebuild_test 00:30:34.139 ************************************ 00:30:34.139 13:50:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:30:34.139 00:30:34.139 real 0m13.913s 00:30:34.139 user 0m15.070s 00:30:34.139 sys 0m2.810s 00:30:34.139 13:50:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:30:34.139 13:50:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:34.140 13:50:40 bdev_raid -- bdev/bdev_raid.sh@979 -- # run_test raid_rebuild_test_sb raid_rebuild_test raid1 2 true false true 00:30:34.140 13:50:40 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:30:34.140 13:50:40 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:30:34.140 13:50:40 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:30:34.140 ************************************ 00:30:34.140 START TEST raid_rebuild_test_sb 00:30:34.140 ************************************ 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false true 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=73609 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 73609 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 73609 ']' 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:30:34.140 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:30:34.140 13:50:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:34.140 [2024-11-20 13:50:41.047867] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:30:34.140 I/O size of 3145728 is greater than zero copy threshold (65536). 00:30:34.140 Zero copy mechanism will not be used. 00:30:34.140 [2024-11-20 13:50:41.048369] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73609 ] 00:30:34.398 [2024-11-20 13:50:41.201796] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:30:34.398 [2024-11-20 13:50:41.305859] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:30:34.398 [2024-11-20 13:50:41.429007] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:30:34.398 [2024-11-20 13:50:41.429056] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:30:34.965 13:50:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:30:34.965 13:50:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:30:34.965 13:50:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:30:34.965 13:50:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:30:34.965 13:50:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:34.965 13:50:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:34.965 BaseBdev1_malloc 00:30:34.965 13:50:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:34.965 13:50:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:30:34.965 13:50:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:34.965 13:50:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:34.965 [2024-11-20 13:50:41.944067] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:30:34.965 [2024-11-20 13:50:41.944149] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:34.965 [2024-11-20 13:50:41.944172] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:30:34.965 [2024-11-20 13:50:41.944183] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:34.965 [2024-11-20 13:50:41.946301] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:34.965 [2024-11-20 13:50:41.946348] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:30:34.965 BaseBdev1 00:30:34.965 13:50:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:34.965 13:50:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:30:34.965 13:50:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:30:34.965 13:50:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:34.965 13:50:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:34.965 BaseBdev2_malloc 00:30:34.965 13:50:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:34.965 13:50:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:30:34.965 13:50:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:34.965 13:50:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:34.965 [2024-11-20 13:50:41.982537] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:30:34.965 [2024-11-20 13:50:41.982614] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:34.965 [2024-11-20 13:50:41.982636] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:30:34.965 [2024-11-20 13:50:41.982647] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:34.965 [2024-11-20 13:50:41.984678] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:34.965 [2024-11-20 13:50:41.984717] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:30:34.965 BaseBdev2 00:30:34.965 13:50:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:34.965 13:50:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:30:34.965 13:50:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:34.965 13:50:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:35.223 spare_malloc 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:35.223 spare_delay 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:35.223 [2024-11-20 13:50:42.047628] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:30:35.223 [2024-11-20 13:50:42.047698] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:35.223 [2024-11-20 13:50:42.047722] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:30:35.223 [2024-11-20 13:50:42.047734] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:35.223 [2024-11-20 13:50:42.049807] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:35.223 [2024-11-20 13:50:42.049847] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:30:35.223 spare 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:35.223 [2024-11-20 13:50:42.059711] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:30:35.223 [2024-11-20 13:50:42.061500] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:30:35.223 [2024-11-20 13:50:42.061676] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:30:35.223 [2024-11-20 13:50:42.061696] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:30:35.223 [2024-11-20 13:50:42.061965] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:30:35.223 [2024-11-20 13:50:42.062136] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:30:35.223 [2024-11-20 13:50:42.062151] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:30:35.223 [2024-11-20 13:50:42.062315] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:35.223 "name": "raid_bdev1", 00:30:35.223 "uuid": "2708905e-0f08-48a5-b125-0c3680528641", 00:30:35.223 "strip_size_kb": 0, 00:30:35.223 "state": "online", 00:30:35.223 "raid_level": "raid1", 00:30:35.223 "superblock": true, 00:30:35.223 "num_base_bdevs": 2, 00:30:35.223 "num_base_bdevs_discovered": 2, 00:30:35.223 "num_base_bdevs_operational": 2, 00:30:35.223 "base_bdevs_list": [ 00:30:35.223 { 00:30:35.223 "name": "BaseBdev1", 00:30:35.223 "uuid": "b31627a2-e034-52bb-a450-dcc696d2f8b2", 00:30:35.223 "is_configured": true, 00:30:35.223 "data_offset": 2048, 00:30:35.223 "data_size": 63488 00:30:35.223 }, 00:30:35.223 { 00:30:35.223 "name": "BaseBdev2", 00:30:35.223 "uuid": "db5686ea-1ece-5d2b-8499-9357b69db901", 00:30:35.223 "is_configured": true, 00:30:35.223 "data_offset": 2048, 00:30:35.223 "data_size": 63488 00:30:35.223 } 00:30:35.223 ] 00:30:35.223 }' 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:35.223 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:35.482 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:30:35.482 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:30:35.482 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:35.482 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:35.482 [2024-11-20 13:50:42.400022] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:30:35.482 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:35.482 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:30:35.482 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:30:35.482 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:35.482 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:35.482 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:35.482 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:35.482 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:30:35.482 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:30:35.482 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:30:35.482 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:30:35.482 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:30:35.482 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:30:35.482 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:30:35.482 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:30:35.482 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:30:35.482 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:30:35.482 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:30:35.482 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:30:35.482 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:30:35.482 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:30:35.744 [2024-11-20 13:50:42.655832] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:30:35.744 /dev/nbd0 00:30:35.744 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:30:35.744 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:30:35.744 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:30:35.744 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:30:35.744 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:30:35.744 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:30:35.744 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:30:35.744 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:30:35.744 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:30:35.744 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:30:35.744 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:30:35.744 1+0 records in 00:30:35.744 1+0 records out 00:30:35.744 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000291947 s, 14.0 MB/s 00:30:35.744 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:30:35.744 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:30:35.744 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:30:35.744 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:30:35.744 13:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:30:35.744 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:30:35.744 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:30:35.744 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:30:35.744 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:30:35.744 13:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=63488 oflag=direct 00:30:41.008 63488+0 records in 00:30:41.008 63488+0 records out 00:30:41.008 32505856 bytes (33 MB, 31 MiB) copied, 4.51305 s, 7.2 MB/s 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:30:41.008 [2024-11-20 13:50:47.408228] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:41.008 [2024-11-20 13:50:47.417555] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:41.008 "name": "raid_bdev1", 00:30:41.008 "uuid": "2708905e-0f08-48a5-b125-0c3680528641", 00:30:41.008 "strip_size_kb": 0, 00:30:41.008 "state": "online", 00:30:41.008 "raid_level": "raid1", 00:30:41.008 "superblock": true, 00:30:41.008 "num_base_bdevs": 2, 00:30:41.008 "num_base_bdevs_discovered": 1, 00:30:41.008 "num_base_bdevs_operational": 1, 00:30:41.008 "base_bdevs_list": [ 00:30:41.008 { 00:30:41.008 "name": null, 00:30:41.008 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:41.008 "is_configured": false, 00:30:41.008 "data_offset": 0, 00:30:41.008 "data_size": 63488 00:30:41.008 }, 00:30:41.008 { 00:30:41.008 "name": "BaseBdev2", 00:30:41.008 "uuid": "db5686ea-1ece-5d2b-8499-9357b69db901", 00:30:41.008 "is_configured": true, 00:30:41.008 "data_offset": 2048, 00:30:41.008 "data_size": 63488 00:30:41.008 } 00:30:41.008 ] 00:30:41.008 }' 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:41.008 [2024-11-20 13:50:47.729663] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:30:41.008 [2024-11-20 13:50:47.740112] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000ca3360 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:41.008 13:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:30:41.008 [2024-11-20 13:50:47.741919] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:30:41.715 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:30:41.715 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:41.715 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:30:41.715 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:30:41.715 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:41.715 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:41.715 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:41.715 13:50:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:41.715 13:50:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:41.715 13:50:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:41.972 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:41.972 "name": "raid_bdev1", 00:30:41.972 "uuid": "2708905e-0f08-48a5-b125-0c3680528641", 00:30:41.972 "strip_size_kb": 0, 00:30:41.972 "state": "online", 00:30:41.972 "raid_level": "raid1", 00:30:41.972 "superblock": true, 00:30:41.972 "num_base_bdevs": 2, 00:30:41.972 "num_base_bdevs_discovered": 2, 00:30:41.972 "num_base_bdevs_operational": 2, 00:30:41.972 "process": { 00:30:41.972 "type": "rebuild", 00:30:41.972 "target": "spare", 00:30:41.972 "progress": { 00:30:41.972 "blocks": 20480, 00:30:41.972 "percent": 32 00:30:41.972 } 00:30:41.972 }, 00:30:41.972 "base_bdevs_list": [ 00:30:41.972 { 00:30:41.972 "name": "spare", 00:30:41.972 "uuid": "cfdd2958-b6ba-5fdf-abd3-33589c91932b", 00:30:41.972 "is_configured": true, 00:30:41.972 "data_offset": 2048, 00:30:41.972 "data_size": 63488 00:30:41.972 }, 00:30:41.972 { 00:30:41.972 "name": "BaseBdev2", 00:30:41.972 "uuid": "db5686ea-1ece-5d2b-8499-9357b69db901", 00:30:41.972 "is_configured": true, 00:30:41.972 "data_offset": 2048, 00:30:41.972 "data_size": 63488 00:30:41.972 } 00:30:41.972 ] 00:30:41.972 }' 00:30:41.972 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:41.972 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:30:41.972 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:41.972 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:30:41.972 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:30:41.972 13:50:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:41.972 13:50:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:41.972 [2024-11-20 13:50:48.843797] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:30:41.972 [2024-11-20 13:50:48.848895] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:30:41.972 [2024-11-20 13:50:48.848973] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:41.972 [2024-11-20 13:50:48.849001] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:30:41.972 [2024-11-20 13:50:48.849011] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:30:41.972 13:50:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:41.972 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:30:41.972 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:41.972 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:41.972 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:41.972 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:41.972 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:30:41.972 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:41.972 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:41.972 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:41.972 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:41.972 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:41.972 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:41.972 13:50:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:41.972 13:50:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:41.972 13:50:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:41.972 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:41.973 "name": "raid_bdev1", 00:30:41.973 "uuid": "2708905e-0f08-48a5-b125-0c3680528641", 00:30:41.973 "strip_size_kb": 0, 00:30:41.973 "state": "online", 00:30:41.973 "raid_level": "raid1", 00:30:41.973 "superblock": true, 00:30:41.973 "num_base_bdevs": 2, 00:30:41.973 "num_base_bdevs_discovered": 1, 00:30:41.973 "num_base_bdevs_operational": 1, 00:30:41.973 "base_bdevs_list": [ 00:30:41.973 { 00:30:41.973 "name": null, 00:30:41.973 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:41.973 "is_configured": false, 00:30:41.973 "data_offset": 0, 00:30:41.973 "data_size": 63488 00:30:41.973 }, 00:30:41.973 { 00:30:41.973 "name": "BaseBdev2", 00:30:41.973 "uuid": "db5686ea-1ece-5d2b-8499-9357b69db901", 00:30:41.973 "is_configured": true, 00:30:41.973 "data_offset": 2048, 00:30:41.973 "data_size": 63488 00:30:41.973 } 00:30:41.973 ] 00:30:41.973 }' 00:30:41.973 13:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:41.973 13:50:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:42.231 13:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:30:42.231 13:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:42.231 13:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:30:42.231 13:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:30:42.231 13:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:42.231 13:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:42.231 13:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:42.231 13:50:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:42.231 13:50:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:42.231 13:50:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:42.231 13:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:42.231 "name": "raid_bdev1", 00:30:42.231 "uuid": "2708905e-0f08-48a5-b125-0c3680528641", 00:30:42.231 "strip_size_kb": 0, 00:30:42.231 "state": "online", 00:30:42.231 "raid_level": "raid1", 00:30:42.231 "superblock": true, 00:30:42.231 "num_base_bdevs": 2, 00:30:42.231 "num_base_bdevs_discovered": 1, 00:30:42.231 "num_base_bdevs_operational": 1, 00:30:42.231 "base_bdevs_list": [ 00:30:42.231 { 00:30:42.231 "name": null, 00:30:42.231 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:42.231 "is_configured": false, 00:30:42.231 "data_offset": 0, 00:30:42.231 "data_size": 63488 00:30:42.231 }, 00:30:42.231 { 00:30:42.231 "name": "BaseBdev2", 00:30:42.231 "uuid": "db5686ea-1ece-5d2b-8499-9357b69db901", 00:30:42.231 "is_configured": true, 00:30:42.231 "data_offset": 2048, 00:30:42.231 "data_size": 63488 00:30:42.231 } 00:30:42.231 ] 00:30:42.231 }' 00:30:42.231 13:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:42.231 13:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:30:42.231 13:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:42.231 13:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:30:42.231 13:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:30:42.231 13:50:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:42.231 13:50:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:42.231 [2024-11-20 13:50:49.249539] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:30:42.231 [2024-11-20 13:50:49.259406] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000ca3430 00:30:42.231 13:50:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:42.231 13:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:30:42.231 [2024-11-20 13:50:49.261186] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:43.604 "name": "raid_bdev1", 00:30:43.604 "uuid": "2708905e-0f08-48a5-b125-0c3680528641", 00:30:43.604 "strip_size_kb": 0, 00:30:43.604 "state": "online", 00:30:43.604 "raid_level": "raid1", 00:30:43.604 "superblock": true, 00:30:43.604 "num_base_bdevs": 2, 00:30:43.604 "num_base_bdevs_discovered": 2, 00:30:43.604 "num_base_bdevs_operational": 2, 00:30:43.604 "process": { 00:30:43.604 "type": "rebuild", 00:30:43.604 "target": "spare", 00:30:43.604 "progress": { 00:30:43.604 "blocks": 20480, 00:30:43.604 "percent": 32 00:30:43.604 } 00:30:43.604 }, 00:30:43.604 "base_bdevs_list": [ 00:30:43.604 { 00:30:43.604 "name": "spare", 00:30:43.604 "uuid": "cfdd2958-b6ba-5fdf-abd3-33589c91932b", 00:30:43.604 "is_configured": true, 00:30:43.604 "data_offset": 2048, 00:30:43.604 "data_size": 63488 00:30:43.604 }, 00:30:43.604 { 00:30:43.604 "name": "BaseBdev2", 00:30:43.604 "uuid": "db5686ea-1ece-5d2b-8499-9357b69db901", 00:30:43.604 "is_configured": true, 00:30:43.604 "data_offset": 2048, 00:30:43.604 "data_size": 63488 00:30:43.604 } 00:30:43.604 ] 00:30:43.604 }' 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:30:43.604 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=297 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:43.604 "name": "raid_bdev1", 00:30:43.604 "uuid": "2708905e-0f08-48a5-b125-0c3680528641", 00:30:43.604 "strip_size_kb": 0, 00:30:43.604 "state": "online", 00:30:43.604 "raid_level": "raid1", 00:30:43.604 "superblock": true, 00:30:43.604 "num_base_bdevs": 2, 00:30:43.604 "num_base_bdevs_discovered": 2, 00:30:43.604 "num_base_bdevs_operational": 2, 00:30:43.604 "process": { 00:30:43.604 "type": "rebuild", 00:30:43.604 "target": "spare", 00:30:43.604 "progress": { 00:30:43.604 "blocks": 22528, 00:30:43.604 "percent": 35 00:30:43.604 } 00:30:43.604 }, 00:30:43.604 "base_bdevs_list": [ 00:30:43.604 { 00:30:43.604 "name": "spare", 00:30:43.604 "uuid": "cfdd2958-b6ba-5fdf-abd3-33589c91932b", 00:30:43.604 "is_configured": true, 00:30:43.604 "data_offset": 2048, 00:30:43.604 "data_size": 63488 00:30:43.604 }, 00:30:43.604 { 00:30:43.604 "name": "BaseBdev2", 00:30:43.604 "uuid": "db5686ea-1ece-5d2b-8499-9357b69db901", 00:30:43.604 "is_configured": true, 00:30:43.604 "data_offset": 2048, 00:30:43.604 "data_size": 63488 00:30:43.604 } 00:30:43.604 ] 00:30:43.604 }' 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:30:43.604 13:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:30:44.542 13:50:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:30:44.542 13:50:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:30:44.542 13:50:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:44.542 13:50:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:30:44.542 13:50:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:30:44.542 13:50:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:44.542 13:50:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:44.542 13:50:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:44.542 13:50:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:44.542 13:50:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:44.542 13:50:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:44.542 13:50:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:44.542 "name": "raid_bdev1", 00:30:44.542 "uuid": "2708905e-0f08-48a5-b125-0c3680528641", 00:30:44.542 "strip_size_kb": 0, 00:30:44.542 "state": "online", 00:30:44.542 "raid_level": "raid1", 00:30:44.542 "superblock": true, 00:30:44.542 "num_base_bdevs": 2, 00:30:44.542 "num_base_bdevs_discovered": 2, 00:30:44.542 "num_base_bdevs_operational": 2, 00:30:44.542 "process": { 00:30:44.542 "type": "rebuild", 00:30:44.542 "target": "spare", 00:30:44.542 "progress": { 00:30:44.542 "blocks": 45056, 00:30:44.542 "percent": 70 00:30:44.542 } 00:30:44.542 }, 00:30:44.542 "base_bdevs_list": [ 00:30:44.542 { 00:30:44.542 "name": "spare", 00:30:44.542 "uuid": "cfdd2958-b6ba-5fdf-abd3-33589c91932b", 00:30:44.542 "is_configured": true, 00:30:44.542 "data_offset": 2048, 00:30:44.542 "data_size": 63488 00:30:44.542 }, 00:30:44.542 { 00:30:44.542 "name": "BaseBdev2", 00:30:44.542 "uuid": "db5686ea-1ece-5d2b-8499-9357b69db901", 00:30:44.542 "is_configured": true, 00:30:44.542 "data_offset": 2048, 00:30:44.542 "data_size": 63488 00:30:44.542 } 00:30:44.542 ] 00:30:44.542 }' 00:30:44.542 13:50:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:44.542 13:50:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:30:44.542 13:50:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:44.542 13:50:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:30:44.542 13:50:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:30:45.476 [2024-11-20 13:50:52.379539] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:30:45.476 [2024-11-20 13:50:52.379639] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:30:45.477 [2024-11-20 13:50:52.379760] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:45.735 "name": "raid_bdev1", 00:30:45.735 "uuid": "2708905e-0f08-48a5-b125-0c3680528641", 00:30:45.735 "strip_size_kb": 0, 00:30:45.735 "state": "online", 00:30:45.735 "raid_level": "raid1", 00:30:45.735 "superblock": true, 00:30:45.735 "num_base_bdevs": 2, 00:30:45.735 "num_base_bdevs_discovered": 2, 00:30:45.735 "num_base_bdevs_operational": 2, 00:30:45.735 "base_bdevs_list": [ 00:30:45.735 { 00:30:45.735 "name": "spare", 00:30:45.735 "uuid": "cfdd2958-b6ba-5fdf-abd3-33589c91932b", 00:30:45.735 "is_configured": true, 00:30:45.735 "data_offset": 2048, 00:30:45.735 "data_size": 63488 00:30:45.735 }, 00:30:45.735 { 00:30:45.735 "name": "BaseBdev2", 00:30:45.735 "uuid": "db5686ea-1ece-5d2b-8499-9357b69db901", 00:30:45.735 "is_configured": true, 00:30:45.735 "data_offset": 2048, 00:30:45.735 "data_size": 63488 00:30:45.735 } 00:30:45.735 ] 00:30:45.735 }' 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:45.735 "name": "raid_bdev1", 00:30:45.735 "uuid": "2708905e-0f08-48a5-b125-0c3680528641", 00:30:45.735 "strip_size_kb": 0, 00:30:45.735 "state": "online", 00:30:45.735 "raid_level": "raid1", 00:30:45.735 "superblock": true, 00:30:45.735 "num_base_bdevs": 2, 00:30:45.735 "num_base_bdevs_discovered": 2, 00:30:45.735 "num_base_bdevs_operational": 2, 00:30:45.735 "base_bdevs_list": [ 00:30:45.735 { 00:30:45.735 "name": "spare", 00:30:45.735 "uuid": "cfdd2958-b6ba-5fdf-abd3-33589c91932b", 00:30:45.735 "is_configured": true, 00:30:45.735 "data_offset": 2048, 00:30:45.735 "data_size": 63488 00:30:45.735 }, 00:30:45.735 { 00:30:45.735 "name": "BaseBdev2", 00:30:45.735 "uuid": "db5686ea-1ece-5d2b-8499-9357b69db901", 00:30:45.735 "is_configured": true, 00:30:45.735 "data_offset": 2048, 00:30:45.735 "data_size": 63488 00:30:45.735 } 00:30:45.735 ] 00:30:45.735 }' 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:45.735 13:50:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:45.994 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:45.994 "name": "raid_bdev1", 00:30:45.994 "uuid": "2708905e-0f08-48a5-b125-0c3680528641", 00:30:45.994 "strip_size_kb": 0, 00:30:45.994 "state": "online", 00:30:45.994 "raid_level": "raid1", 00:30:45.994 "superblock": true, 00:30:45.994 "num_base_bdevs": 2, 00:30:45.994 "num_base_bdevs_discovered": 2, 00:30:45.994 "num_base_bdevs_operational": 2, 00:30:45.994 "base_bdevs_list": [ 00:30:45.994 { 00:30:45.994 "name": "spare", 00:30:45.994 "uuid": "cfdd2958-b6ba-5fdf-abd3-33589c91932b", 00:30:45.994 "is_configured": true, 00:30:45.994 "data_offset": 2048, 00:30:45.994 "data_size": 63488 00:30:45.994 }, 00:30:45.994 { 00:30:45.994 "name": "BaseBdev2", 00:30:45.994 "uuid": "db5686ea-1ece-5d2b-8499-9357b69db901", 00:30:45.994 "is_configured": true, 00:30:45.994 "data_offset": 2048, 00:30:45.994 "data_size": 63488 00:30:45.994 } 00:30:45.994 ] 00:30:45.994 }' 00:30:45.994 13:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:45.994 13:50:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:46.252 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:30:46.252 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:46.252 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:46.252 [2024-11-20 13:50:53.103395] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:30:46.252 [2024-11-20 13:50:53.103434] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:30:46.252 [2024-11-20 13:50:53.103512] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:30:46.252 [2024-11-20 13:50:53.103578] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:30:46.252 [2024-11-20 13:50:53.103593] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:30:46.252 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:46.252 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:46.252 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:46.252 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:30:46.252 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:46.252 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:46.252 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:30:46.252 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:30:46.252 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:30:46.252 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:30:46.252 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:30:46.252 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:30:46.252 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:30:46.252 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:30:46.252 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:30:46.252 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:30:46.252 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:30:46.252 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:30:46.252 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:30:46.511 /dev/nbd0 00:30:46.511 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:30:46.511 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:30:46.511 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:30:46.511 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:30:46.511 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:30:46.511 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:30:46.511 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:30:46.511 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:30:46.511 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:30:46.511 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:30:46.511 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:30:46.511 1+0 records in 00:30:46.511 1+0 records out 00:30:46.511 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000207277 s, 19.8 MB/s 00:30:46.511 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:30:46.511 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:30:46.511 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:30:46.511 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:30:46.511 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:30:46.511 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:30:46.511 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:30:46.511 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:30:46.770 /dev/nbd1 00:30:46.770 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:30:46.770 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:30:46.770 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:30:46.770 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:30:46.770 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:30:46.770 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:30:46.770 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:30:46.770 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:30:46.770 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:30:46.770 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:30:46.770 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:30:46.770 1+0 records in 00:30:46.770 1+0 records out 00:30:46.770 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000352475 s, 11.6 MB/s 00:30:46.770 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:30:46.770 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:30:46.770 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:30:46.770 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:30:46.770 13:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:30:46.770 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:30:46.770 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:30:46.770 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:30:46.770 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:30:46.770 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:30:46.770 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:30:46.770 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:30:46.770 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:30:46.770 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:30:46.770 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:30:47.027 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:30:47.027 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:30:47.027 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:30:47.027 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:30:47.027 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:30:47.027 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:30:47.027 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:30:47.027 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:30:47.027 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:30:47.027 13:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:30:47.285 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:30:47.285 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:30:47.285 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:47.286 [2024-11-20 13:50:54.168897] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:30:47.286 [2024-11-20 13:50:54.168953] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:47.286 [2024-11-20 13:50:54.168988] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:30:47.286 [2024-11-20 13:50:54.168997] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:47.286 [2024-11-20 13:50:54.171010] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:47.286 [2024-11-20 13:50:54.171041] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:30:47.286 [2024-11-20 13:50:54.171127] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:30:47.286 [2024-11-20 13:50:54.171169] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:30:47.286 [2024-11-20 13:50:54.171288] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:30:47.286 spare 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:47.286 [2024-11-20 13:50:54.271382] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:30:47.286 [2024-11-20 13:50:54.271433] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:30:47.286 [2024-11-20 13:50:54.271750] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1ae0 00:30:47.286 [2024-11-20 13:50:54.271940] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:30:47.286 [2024-11-20 13:50:54.271958] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:30:47.286 [2024-11-20 13:50:54.272134] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:47.286 "name": "raid_bdev1", 00:30:47.286 "uuid": "2708905e-0f08-48a5-b125-0c3680528641", 00:30:47.286 "strip_size_kb": 0, 00:30:47.286 "state": "online", 00:30:47.286 "raid_level": "raid1", 00:30:47.286 "superblock": true, 00:30:47.286 "num_base_bdevs": 2, 00:30:47.286 "num_base_bdevs_discovered": 2, 00:30:47.286 "num_base_bdevs_operational": 2, 00:30:47.286 "base_bdevs_list": [ 00:30:47.286 { 00:30:47.286 "name": "spare", 00:30:47.286 "uuid": "cfdd2958-b6ba-5fdf-abd3-33589c91932b", 00:30:47.286 "is_configured": true, 00:30:47.286 "data_offset": 2048, 00:30:47.286 "data_size": 63488 00:30:47.286 }, 00:30:47.286 { 00:30:47.286 "name": "BaseBdev2", 00:30:47.286 "uuid": "db5686ea-1ece-5d2b-8499-9357b69db901", 00:30:47.286 "is_configured": true, 00:30:47.286 "data_offset": 2048, 00:30:47.286 "data_size": 63488 00:30:47.286 } 00:30:47.286 ] 00:30:47.286 }' 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:47.286 13:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:47.544 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:30:47.544 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:47.544 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:30:47.544 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:30:47.544 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:47.544 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:47.544 13:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:47.544 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:47.544 13:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:47.802 "name": "raid_bdev1", 00:30:47.802 "uuid": "2708905e-0f08-48a5-b125-0c3680528641", 00:30:47.802 "strip_size_kb": 0, 00:30:47.802 "state": "online", 00:30:47.802 "raid_level": "raid1", 00:30:47.802 "superblock": true, 00:30:47.802 "num_base_bdevs": 2, 00:30:47.802 "num_base_bdevs_discovered": 2, 00:30:47.802 "num_base_bdevs_operational": 2, 00:30:47.802 "base_bdevs_list": [ 00:30:47.802 { 00:30:47.802 "name": "spare", 00:30:47.802 "uuid": "cfdd2958-b6ba-5fdf-abd3-33589c91932b", 00:30:47.802 "is_configured": true, 00:30:47.802 "data_offset": 2048, 00:30:47.802 "data_size": 63488 00:30:47.802 }, 00:30:47.802 { 00:30:47.802 "name": "BaseBdev2", 00:30:47.802 "uuid": "db5686ea-1ece-5d2b-8499-9357b69db901", 00:30:47.802 "is_configured": true, 00:30:47.802 "data_offset": 2048, 00:30:47.802 "data_size": 63488 00:30:47.802 } 00:30:47.802 ] 00:30:47.802 }' 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:47.802 [2024-11-20 13:50:54.721077] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:47.802 "name": "raid_bdev1", 00:30:47.802 "uuid": "2708905e-0f08-48a5-b125-0c3680528641", 00:30:47.802 "strip_size_kb": 0, 00:30:47.802 "state": "online", 00:30:47.802 "raid_level": "raid1", 00:30:47.802 "superblock": true, 00:30:47.802 "num_base_bdevs": 2, 00:30:47.802 "num_base_bdevs_discovered": 1, 00:30:47.802 "num_base_bdevs_operational": 1, 00:30:47.802 "base_bdevs_list": [ 00:30:47.802 { 00:30:47.802 "name": null, 00:30:47.802 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:47.802 "is_configured": false, 00:30:47.802 "data_offset": 0, 00:30:47.802 "data_size": 63488 00:30:47.802 }, 00:30:47.802 { 00:30:47.802 "name": "BaseBdev2", 00:30:47.802 "uuid": "db5686ea-1ece-5d2b-8499-9357b69db901", 00:30:47.802 "is_configured": true, 00:30:47.802 "data_offset": 2048, 00:30:47.802 "data_size": 63488 00:30:47.802 } 00:30:47.802 ] 00:30:47.802 }' 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:47.802 13:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:48.060 13:50:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:30:48.060 13:50:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:48.060 13:50:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:48.060 [2024-11-20 13:50:55.109174] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:30:48.060 [2024-11-20 13:50:55.109372] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:30:48.060 [2024-11-20 13:50:55.109387] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:30:48.060 [2024-11-20 13:50:55.109425] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:30:48.318 [2024-11-20 13:50:55.118952] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1bb0 00:30:48.318 13:50:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:48.318 13:50:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:30:48.318 [2024-11-20 13:50:55.120704] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:30:49.250 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:30:49.250 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:49.250 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:30:49.250 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:30:49.250 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:49.250 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:49.250 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:49.250 13:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:49.250 13:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:49.250 13:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:49.250 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:49.250 "name": "raid_bdev1", 00:30:49.250 "uuid": "2708905e-0f08-48a5-b125-0c3680528641", 00:30:49.250 "strip_size_kb": 0, 00:30:49.250 "state": "online", 00:30:49.250 "raid_level": "raid1", 00:30:49.250 "superblock": true, 00:30:49.250 "num_base_bdevs": 2, 00:30:49.250 "num_base_bdevs_discovered": 2, 00:30:49.250 "num_base_bdevs_operational": 2, 00:30:49.250 "process": { 00:30:49.250 "type": "rebuild", 00:30:49.250 "target": "spare", 00:30:49.250 "progress": { 00:30:49.250 "blocks": 20480, 00:30:49.250 "percent": 32 00:30:49.250 } 00:30:49.250 }, 00:30:49.250 "base_bdevs_list": [ 00:30:49.250 { 00:30:49.250 "name": "spare", 00:30:49.250 "uuid": "cfdd2958-b6ba-5fdf-abd3-33589c91932b", 00:30:49.250 "is_configured": true, 00:30:49.250 "data_offset": 2048, 00:30:49.250 "data_size": 63488 00:30:49.250 }, 00:30:49.250 { 00:30:49.250 "name": "BaseBdev2", 00:30:49.250 "uuid": "db5686ea-1ece-5d2b-8499-9357b69db901", 00:30:49.250 "is_configured": true, 00:30:49.250 "data_offset": 2048, 00:30:49.250 "data_size": 63488 00:30:49.250 } 00:30:49.250 ] 00:30:49.250 }' 00:30:49.251 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:49.251 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:30:49.251 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:49.251 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:30:49.251 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:30:49.251 13:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:49.251 13:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:49.251 [2024-11-20 13:50:56.230730] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:30:49.508 [2024-11-20 13:50:56.327765] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:30:49.508 [2024-11-20 13:50:56.327849] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:49.508 [2024-11-20 13:50:56.327862] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:30:49.508 [2024-11-20 13:50:56.327870] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:30:49.508 13:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:49.508 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:30:49.508 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:49.508 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:49.508 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:49.508 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:49.508 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:30:49.508 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:49.508 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:49.508 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:49.508 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:49.508 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:49.508 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:49.508 13:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:49.508 13:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:49.508 13:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:49.508 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:49.508 "name": "raid_bdev1", 00:30:49.508 "uuid": "2708905e-0f08-48a5-b125-0c3680528641", 00:30:49.508 "strip_size_kb": 0, 00:30:49.508 "state": "online", 00:30:49.508 "raid_level": "raid1", 00:30:49.508 "superblock": true, 00:30:49.508 "num_base_bdevs": 2, 00:30:49.509 "num_base_bdevs_discovered": 1, 00:30:49.509 "num_base_bdevs_operational": 1, 00:30:49.509 "base_bdevs_list": [ 00:30:49.509 { 00:30:49.509 "name": null, 00:30:49.509 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:49.509 "is_configured": false, 00:30:49.509 "data_offset": 0, 00:30:49.509 "data_size": 63488 00:30:49.509 }, 00:30:49.509 { 00:30:49.509 "name": "BaseBdev2", 00:30:49.509 "uuid": "db5686ea-1ece-5d2b-8499-9357b69db901", 00:30:49.509 "is_configured": true, 00:30:49.509 "data_offset": 2048, 00:30:49.509 "data_size": 63488 00:30:49.509 } 00:30:49.509 ] 00:30:49.509 }' 00:30:49.509 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:49.509 13:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:49.767 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:30:49.767 13:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:49.767 13:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:49.767 [2024-11-20 13:50:56.675499] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:30:49.767 [2024-11-20 13:50:56.675576] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:49.767 [2024-11-20 13:50:56.675597] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:30:49.767 [2024-11-20 13:50:56.675607] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:49.767 [2024-11-20 13:50:56.676060] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:49.767 [2024-11-20 13:50:56.676077] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:30:49.767 [2024-11-20 13:50:56.676164] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:30:49.767 [2024-11-20 13:50:56.676178] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:30:49.767 [2024-11-20 13:50:56.676188] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:30:49.767 [2024-11-20 13:50:56.676212] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:30:49.767 [2024-11-20 13:50:56.685836] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1c80 00:30:49.767 spare 00:30:49.767 13:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:49.767 13:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:30:49.767 [2024-11-20 13:50:56.687662] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:30:50.716 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:30:50.716 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:50.716 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:30:50.716 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:30:50.716 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:50.716 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:50.716 13:50:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:50.716 13:50:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:50.716 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:50.716 13:50:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:50.716 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:50.716 "name": "raid_bdev1", 00:30:50.716 "uuid": "2708905e-0f08-48a5-b125-0c3680528641", 00:30:50.716 "strip_size_kb": 0, 00:30:50.716 "state": "online", 00:30:50.716 "raid_level": "raid1", 00:30:50.716 "superblock": true, 00:30:50.716 "num_base_bdevs": 2, 00:30:50.716 "num_base_bdevs_discovered": 2, 00:30:50.716 "num_base_bdevs_operational": 2, 00:30:50.716 "process": { 00:30:50.716 "type": "rebuild", 00:30:50.716 "target": "spare", 00:30:50.716 "progress": { 00:30:50.716 "blocks": 20480, 00:30:50.716 "percent": 32 00:30:50.716 } 00:30:50.716 }, 00:30:50.716 "base_bdevs_list": [ 00:30:50.716 { 00:30:50.716 "name": "spare", 00:30:50.716 "uuid": "cfdd2958-b6ba-5fdf-abd3-33589c91932b", 00:30:50.716 "is_configured": true, 00:30:50.716 "data_offset": 2048, 00:30:50.716 "data_size": 63488 00:30:50.716 }, 00:30:50.716 { 00:30:50.716 "name": "BaseBdev2", 00:30:50.716 "uuid": "db5686ea-1ece-5d2b-8499-9357b69db901", 00:30:50.716 "is_configured": true, 00:30:50.716 "data_offset": 2048, 00:30:50.716 "data_size": 63488 00:30:50.716 } 00:30:50.716 ] 00:30:50.716 }' 00:30:50.716 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:50.716 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:30:50.977 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:50.977 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:30:50.977 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:30:50.977 13:50:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:50.977 13:50:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:50.977 [2024-11-20 13:50:57.809926] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:30:50.977 [2024-11-20 13:50:57.894993] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:30:50.977 [2024-11-20 13:50:57.895073] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:50.977 [2024-11-20 13:50:57.895089] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:30:50.977 [2024-11-20 13:50:57.895096] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:30:50.977 13:50:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:50.977 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:30:50.977 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:50.977 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:50.977 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:50.977 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:50.977 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:30:50.978 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:50.978 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:50.978 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:50.978 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:50.978 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:50.978 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:50.978 13:50:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:50.978 13:50:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:50.978 13:50:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:50.978 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:50.978 "name": "raid_bdev1", 00:30:50.978 "uuid": "2708905e-0f08-48a5-b125-0c3680528641", 00:30:50.978 "strip_size_kb": 0, 00:30:50.978 "state": "online", 00:30:50.978 "raid_level": "raid1", 00:30:50.978 "superblock": true, 00:30:50.978 "num_base_bdevs": 2, 00:30:50.978 "num_base_bdevs_discovered": 1, 00:30:50.978 "num_base_bdevs_operational": 1, 00:30:50.978 "base_bdevs_list": [ 00:30:50.978 { 00:30:50.978 "name": null, 00:30:50.978 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:50.978 "is_configured": false, 00:30:50.978 "data_offset": 0, 00:30:50.978 "data_size": 63488 00:30:50.978 }, 00:30:50.978 { 00:30:50.978 "name": "BaseBdev2", 00:30:50.978 "uuid": "db5686ea-1ece-5d2b-8499-9357b69db901", 00:30:50.978 "is_configured": true, 00:30:50.978 "data_offset": 2048, 00:30:50.978 "data_size": 63488 00:30:50.978 } 00:30:50.978 ] 00:30:50.978 }' 00:30:50.978 13:50:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:50.978 13:50:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:51.236 13:50:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:30:51.236 13:50:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:51.236 13:50:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:30:51.236 13:50:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:30:51.236 13:50:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:51.236 13:50:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:51.236 13:50:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:51.236 13:50:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:51.236 13:50:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:51.236 13:50:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:51.236 13:50:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:51.236 "name": "raid_bdev1", 00:30:51.236 "uuid": "2708905e-0f08-48a5-b125-0c3680528641", 00:30:51.236 "strip_size_kb": 0, 00:30:51.236 "state": "online", 00:30:51.237 "raid_level": "raid1", 00:30:51.237 "superblock": true, 00:30:51.237 "num_base_bdevs": 2, 00:30:51.237 "num_base_bdevs_discovered": 1, 00:30:51.237 "num_base_bdevs_operational": 1, 00:30:51.237 "base_bdevs_list": [ 00:30:51.237 { 00:30:51.237 "name": null, 00:30:51.237 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:51.237 "is_configured": false, 00:30:51.237 "data_offset": 0, 00:30:51.237 "data_size": 63488 00:30:51.237 }, 00:30:51.237 { 00:30:51.237 "name": "BaseBdev2", 00:30:51.237 "uuid": "db5686ea-1ece-5d2b-8499-9357b69db901", 00:30:51.237 "is_configured": true, 00:30:51.237 "data_offset": 2048, 00:30:51.237 "data_size": 63488 00:30:51.237 } 00:30:51.237 ] 00:30:51.237 }' 00:30:51.237 13:50:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:51.237 13:50:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:30:51.237 13:50:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:51.496 13:50:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:30:51.496 13:50:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:30:51.496 13:50:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:51.496 13:50:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:51.496 13:50:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:51.496 13:50:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:30:51.496 13:50:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:51.496 13:50:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:51.496 [2024-11-20 13:50:58.326999] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:30:51.496 [2024-11-20 13:50:58.327059] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:51.496 [2024-11-20 13:50:58.327082] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:30:51.496 [2024-11-20 13:50:58.327091] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:51.496 [2024-11-20 13:50:58.327511] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:51.496 [2024-11-20 13:50:58.327523] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:30:51.496 [2024-11-20 13:50:58.327595] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:30:51.496 [2024-11-20 13:50:58.327608] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:30:51.496 [2024-11-20 13:50:58.327616] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:30:51.496 [2024-11-20 13:50:58.327625] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:30:51.496 BaseBdev1 00:30:51.496 13:50:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:51.496 13:50:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:30:52.431 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:30:52.431 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:52.431 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:52.431 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:52.431 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:52.431 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:30:52.431 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:52.431 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:52.431 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:52.431 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:52.431 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:52.431 13:50:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:52.431 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:52.431 13:50:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:52.431 13:50:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:52.431 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:52.431 "name": "raid_bdev1", 00:30:52.431 "uuid": "2708905e-0f08-48a5-b125-0c3680528641", 00:30:52.431 "strip_size_kb": 0, 00:30:52.431 "state": "online", 00:30:52.431 "raid_level": "raid1", 00:30:52.431 "superblock": true, 00:30:52.431 "num_base_bdevs": 2, 00:30:52.431 "num_base_bdevs_discovered": 1, 00:30:52.431 "num_base_bdevs_operational": 1, 00:30:52.431 "base_bdevs_list": [ 00:30:52.431 { 00:30:52.431 "name": null, 00:30:52.431 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:52.431 "is_configured": false, 00:30:52.431 "data_offset": 0, 00:30:52.431 "data_size": 63488 00:30:52.431 }, 00:30:52.431 { 00:30:52.431 "name": "BaseBdev2", 00:30:52.431 "uuid": "db5686ea-1ece-5d2b-8499-9357b69db901", 00:30:52.431 "is_configured": true, 00:30:52.431 "data_offset": 2048, 00:30:52.431 "data_size": 63488 00:30:52.431 } 00:30:52.431 ] 00:30:52.431 }' 00:30:52.431 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:52.431 13:50:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:52.691 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:30:52.691 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:52.691 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:30:52.691 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:30:52.691 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:52.691 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:52.691 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:52.691 13:50:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:52.691 13:50:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:52.691 13:50:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:52.691 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:52.691 "name": "raid_bdev1", 00:30:52.691 "uuid": "2708905e-0f08-48a5-b125-0c3680528641", 00:30:52.691 "strip_size_kb": 0, 00:30:52.691 "state": "online", 00:30:52.691 "raid_level": "raid1", 00:30:52.691 "superblock": true, 00:30:52.691 "num_base_bdevs": 2, 00:30:52.691 "num_base_bdevs_discovered": 1, 00:30:52.691 "num_base_bdevs_operational": 1, 00:30:52.691 "base_bdevs_list": [ 00:30:52.691 { 00:30:52.691 "name": null, 00:30:52.691 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:52.691 "is_configured": false, 00:30:52.691 "data_offset": 0, 00:30:52.691 "data_size": 63488 00:30:52.691 }, 00:30:52.691 { 00:30:52.691 "name": "BaseBdev2", 00:30:52.692 "uuid": "db5686ea-1ece-5d2b-8499-9357b69db901", 00:30:52.692 "is_configured": true, 00:30:52.692 "data_offset": 2048, 00:30:52.692 "data_size": 63488 00:30:52.692 } 00:30:52.692 ] 00:30:52.692 }' 00:30:52.692 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:52.692 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:30:52.692 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:52.692 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:30:52.692 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:30:52.692 13:50:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:30:52.692 13:50:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:30:52.692 13:50:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:30:52.692 13:50:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:30:52.692 13:50:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:30:52.692 13:50:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:30:52.692 13:50:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:30:52.692 13:50:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:52.692 13:50:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:52.964 [2024-11-20 13:50:59.747313] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:30:52.964 [2024-11-20 13:50:59.747473] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:30:52.964 [2024-11-20 13:50:59.747488] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:30:52.964 request: 00:30:52.964 { 00:30:52.964 "base_bdev": "BaseBdev1", 00:30:52.964 "raid_bdev": "raid_bdev1", 00:30:52.964 "method": "bdev_raid_add_base_bdev", 00:30:52.964 "req_id": 1 00:30:52.964 } 00:30:52.964 Got JSON-RPC error response 00:30:52.964 response: 00:30:52.964 { 00:30:52.964 "code": -22, 00:30:52.964 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:30:52.964 } 00:30:52.964 13:50:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:30:52.964 13:50:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:30:52.964 13:50:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:30:52.964 13:50:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:30:52.964 13:50:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:30:52.964 13:50:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:30:53.898 13:51:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:30:53.898 13:51:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:53.898 13:51:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:53.898 13:51:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:53.898 13:51:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:53.898 13:51:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:30:53.898 13:51:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:53.898 13:51:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:53.898 13:51:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:53.898 13:51:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:53.898 13:51:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:53.898 13:51:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:53.898 13:51:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:53.898 13:51:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:53.898 13:51:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:53.898 13:51:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:53.898 "name": "raid_bdev1", 00:30:53.898 "uuid": "2708905e-0f08-48a5-b125-0c3680528641", 00:30:53.898 "strip_size_kb": 0, 00:30:53.898 "state": "online", 00:30:53.898 "raid_level": "raid1", 00:30:53.898 "superblock": true, 00:30:53.898 "num_base_bdevs": 2, 00:30:53.898 "num_base_bdevs_discovered": 1, 00:30:53.898 "num_base_bdevs_operational": 1, 00:30:53.898 "base_bdevs_list": [ 00:30:53.898 { 00:30:53.898 "name": null, 00:30:53.898 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:53.898 "is_configured": false, 00:30:53.898 "data_offset": 0, 00:30:53.898 "data_size": 63488 00:30:53.898 }, 00:30:53.898 { 00:30:53.898 "name": "BaseBdev2", 00:30:53.898 "uuid": "db5686ea-1ece-5d2b-8499-9357b69db901", 00:30:53.898 "is_configured": true, 00:30:53.898 "data_offset": 2048, 00:30:53.898 "data_size": 63488 00:30:53.898 } 00:30:53.898 ] 00:30:53.898 }' 00:30:53.898 13:51:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:53.898 13:51:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:54.156 13:51:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:30:54.156 13:51:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:54.156 13:51:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:30:54.156 13:51:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:30:54.156 13:51:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:54.156 13:51:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:54.156 13:51:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:54.156 13:51:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:54.156 13:51:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:54.156 13:51:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:54.156 13:51:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:54.156 "name": "raid_bdev1", 00:30:54.156 "uuid": "2708905e-0f08-48a5-b125-0c3680528641", 00:30:54.156 "strip_size_kb": 0, 00:30:54.156 "state": "online", 00:30:54.156 "raid_level": "raid1", 00:30:54.156 "superblock": true, 00:30:54.156 "num_base_bdevs": 2, 00:30:54.156 "num_base_bdevs_discovered": 1, 00:30:54.156 "num_base_bdevs_operational": 1, 00:30:54.156 "base_bdevs_list": [ 00:30:54.156 { 00:30:54.156 "name": null, 00:30:54.156 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:54.156 "is_configured": false, 00:30:54.156 "data_offset": 0, 00:30:54.156 "data_size": 63488 00:30:54.156 }, 00:30:54.156 { 00:30:54.156 "name": "BaseBdev2", 00:30:54.156 "uuid": "db5686ea-1ece-5d2b-8499-9357b69db901", 00:30:54.156 "is_configured": true, 00:30:54.156 "data_offset": 2048, 00:30:54.156 "data_size": 63488 00:30:54.156 } 00:30:54.156 ] 00:30:54.156 }' 00:30:54.156 13:51:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:54.156 13:51:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:30:54.156 13:51:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:54.156 13:51:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:30:54.156 13:51:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 73609 00:30:54.156 13:51:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 73609 ']' 00:30:54.156 13:51:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 73609 00:30:54.156 13:51:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:30:54.156 13:51:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:30:54.156 13:51:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73609 00:30:54.156 killing process with pid 73609 00:30:54.156 Received shutdown signal, test time was about 60.000000 seconds 00:30:54.156 00:30:54.156 Latency(us) 00:30:54.156 [2024-11-20T13:51:01.215Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:30:54.156 [2024-11-20T13:51:01.215Z] =================================================================================================================== 00:30:54.156 [2024-11-20T13:51:01.215Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:30:54.156 13:51:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:30:54.156 13:51:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:30:54.156 13:51:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73609' 00:30:54.156 13:51:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 73609 00:30:54.156 [2024-11-20 13:51:01.210664] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:30:54.156 13:51:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 73609 00:30:54.156 [2024-11-20 13:51:01.210791] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:30:54.156 [2024-11-20 13:51:01.210841] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:30:54.156 [2024-11-20 13:51:01.210852] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:30:54.414 [2024-11-20 13:51:01.366885] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:30:54.979 ************************************ 00:30:54.979 END TEST raid_rebuild_test_sb 00:30:54.979 ************************************ 00:30:54.979 13:51:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:30:54.979 00:30:54.979 real 0m20.991s 00:30:54.979 user 0m24.569s 00:30:54.979 sys 0m3.253s 00:30:54.979 13:51:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:30:54.979 13:51:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:55.237 13:51:02 bdev_raid -- bdev/bdev_raid.sh@980 -- # run_test raid_rebuild_test_io raid_rebuild_test raid1 2 false true true 00:30:55.237 13:51:02 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:30:55.237 13:51:02 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:30:55.237 13:51:02 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:30:55.237 ************************************ 00:30:55.237 START TEST raid_rebuild_test_io 00:30:55.237 ************************************ 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 false true true 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:30:55.237 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@597 -- # raid_pid=74319 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 74319 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@835 -- # '[' -z 74319 ']' 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:30:55.237 13:51:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:30:55.237 I/O size of 3145728 is greater than zero copy threshold (65536). 00:30:55.237 Zero copy mechanism will not be used. 00:30:55.237 [2024-11-20 13:51:02.141628] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:30:55.237 [2024-11-20 13:51:02.141770] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74319 ] 00:30:55.494 [2024-11-20 13:51:02.301527] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:30:55.494 [2024-11-20 13:51:02.419118] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:30:55.752 [2024-11-20 13:51:02.564933] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:30:55.752 [2024-11-20 13:51:02.564975] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:30:56.010 13:51:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:30:56.010 13:51:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # return 0 00:30:56.010 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:30:56.010 13:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:30:56.010 13:51:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:56.010 13:51:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:30:56.010 BaseBdev1_malloc 00:30:56.010 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:56.010 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:30:56.010 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:56.010 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:30:56.010 [2024-11-20 13:51:03.010819] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:30:56.010 [2024-11-20 13:51:03.011058] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:56.010 [2024-11-20 13:51:03.011089] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:30:56.010 [2024-11-20 13:51:03.011101] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:56.010 [2024-11-20 13:51:03.013345] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:56.010 [2024-11-20 13:51:03.013382] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:30:56.010 BaseBdev1 00:30:56.010 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:56.010 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:30:56.010 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:30:56.010 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:56.010 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:30:56.010 BaseBdev2_malloc 00:30:56.010 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:56.010 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:30:56.010 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:56.010 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:30:56.010 [2024-11-20 13:51:03.048820] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:30:56.010 [2024-11-20 13:51:03.048884] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:56.010 [2024-11-20 13:51:03.048906] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:30:56.010 [2024-11-20 13:51:03.048917] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:56.010 [2024-11-20 13:51:03.051143] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:56.010 [2024-11-20 13:51:03.051177] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:30:56.010 BaseBdev2 00:30:56.010 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:56.010 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:30:56.010 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:56.010 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:30:56.268 spare_malloc 00:30:56.268 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:56.268 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:30:56.268 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:56.268 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:30:56.268 spare_delay 00:30:56.268 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:56.268 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:30:56.268 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:56.268 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:30:56.268 [2024-11-20 13:51:03.111704] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:30:56.268 [2024-11-20 13:51:03.111922] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:56.268 [2024-11-20 13:51:03.111949] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:30:56.268 [2024-11-20 13:51:03.111962] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:56.268 [2024-11-20 13:51:03.114264] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:56.268 [2024-11-20 13:51:03.114305] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:30:56.268 spare 00:30:56.268 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:56.268 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:30:56.268 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:56.268 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:30:56.268 [2024-11-20 13:51:03.119754] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:30:56.268 [2024-11-20 13:51:03.121701] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:30:56.268 [2024-11-20 13:51:03.121791] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:30:56.268 [2024-11-20 13:51:03.121805] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:30:56.268 [2024-11-20 13:51:03.122080] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:30:56.269 [2024-11-20 13:51:03.122232] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:30:56.269 [2024-11-20 13:51:03.122332] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:30:56.269 [2024-11-20 13:51:03.122487] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:56.269 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:56.269 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:30:56.269 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:56.269 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:56.269 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:56.269 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:56.269 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:30:56.269 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:56.269 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:56.269 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:56.269 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:56.269 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:56.269 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:56.269 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:56.269 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:30:56.269 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:56.269 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:56.269 "name": "raid_bdev1", 00:30:56.269 "uuid": "a9479a2c-f98a-4794-8c37-b3989098e8da", 00:30:56.269 "strip_size_kb": 0, 00:30:56.269 "state": "online", 00:30:56.269 "raid_level": "raid1", 00:30:56.269 "superblock": false, 00:30:56.269 "num_base_bdevs": 2, 00:30:56.269 "num_base_bdevs_discovered": 2, 00:30:56.269 "num_base_bdevs_operational": 2, 00:30:56.269 "base_bdevs_list": [ 00:30:56.269 { 00:30:56.269 "name": "BaseBdev1", 00:30:56.269 "uuid": "bfb5615c-73c6-5a8a-9b3c-1034481aabc6", 00:30:56.269 "is_configured": true, 00:30:56.269 "data_offset": 0, 00:30:56.269 "data_size": 65536 00:30:56.269 }, 00:30:56.269 { 00:30:56.269 "name": "BaseBdev2", 00:30:56.269 "uuid": "803f16d6-fd3c-5841-9499-4fcf3fb23aa7", 00:30:56.269 "is_configured": true, 00:30:56.269 "data_offset": 0, 00:30:56.269 "data_size": 65536 00:30:56.269 } 00:30:56.269 ] 00:30:56.269 }' 00:30:56.269 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:56.269 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:30:56.527 [2024-11-20 13:51:03.424167] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:30:56.527 [2024-11-20 13:51:03.491817] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:56.527 "name": "raid_bdev1", 00:30:56.527 "uuid": "a9479a2c-f98a-4794-8c37-b3989098e8da", 00:30:56.527 "strip_size_kb": 0, 00:30:56.527 "state": "online", 00:30:56.527 "raid_level": "raid1", 00:30:56.527 "superblock": false, 00:30:56.527 "num_base_bdevs": 2, 00:30:56.527 "num_base_bdevs_discovered": 1, 00:30:56.527 "num_base_bdevs_operational": 1, 00:30:56.527 "base_bdevs_list": [ 00:30:56.527 { 00:30:56.527 "name": null, 00:30:56.527 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:56.527 "is_configured": false, 00:30:56.527 "data_offset": 0, 00:30:56.527 "data_size": 65536 00:30:56.527 }, 00:30:56.527 { 00:30:56.527 "name": "BaseBdev2", 00:30:56.527 "uuid": "803f16d6-fd3c-5841-9499-4fcf3fb23aa7", 00:30:56.527 "is_configured": true, 00:30:56.527 "data_offset": 0, 00:30:56.527 "data_size": 65536 00:30:56.527 } 00:30:56.527 ] 00:30:56.527 }' 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:56.527 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:30:56.527 [2024-11-20 13:51:03.577625] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:30:56.527 I/O size of 3145728 is greater than zero copy threshold (65536). 00:30:56.527 Zero copy mechanism will not be used. 00:30:56.527 Running I/O for 60 seconds... 00:30:56.785 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:30:56.785 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:56.785 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:30:56.785 [2024-11-20 13:51:03.793250] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:30:56.785 13:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:56.785 13:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:30:57.042 [2024-11-20 13:51:03.850281] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:30:57.042 [2024-11-20 13:51:03.852308] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:30:57.042 [2024-11-20 13:51:03.965887] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:30:57.042 [2024-11-20 13:51:03.966404] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:30:57.607 [2024-11-20 13:51:04.494902] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:30:57.607 151.00 IOPS, 453.00 MiB/s [2024-11-20T13:51:04.666Z] [2024-11-20 13:51:04.625355] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:30:57.607 [2024-11-20 13:51:04.625779] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:30:57.865 13:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:30:57.865 13:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:57.865 13:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:30:57.865 13:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:30:57.865 13:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:57.865 13:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:57.865 13:51:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:57.865 13:51:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:30:57.865 13:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:57.865 [2024-11-20 13:51:04.844725] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:30:57.865 13:51:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:57.865 13:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:57.865 "name": "raid_bdev1", 00:30:57.865 "uuid": "a9479a2c-f98a-4794-8c37-b3989098e8da", 00:30:57.865 "strip_size_kb": 0, 00:30:57.865 "state": "online", 00:30:57.865 "raid_level": "raid1", 00:30:57.865 "superblock": false, 00:30:57.865 "num_base_bdevs": 2, 00:30:57.865 "num_base_bdevs_discovered": 2, 00:30:57.865 "num_base_bdevs_operational": 2, 00:30:57.865 "process": { 00:30:57.865 "type": "rebuild", 00:30:57.865 "target": "spare", 00:30:57.865 "progress": { 00:30:57.865 "blocks": 12288, 00:30:57.865 "percent": 18 00:30:57.865 } 00:30:57.865 }, 00:30:57.865 "base_bdevs_list": [ 00:30:57.865 { 00:30:57.865 "name": "spare", 00:30:57.865 "uuid": "3dd90581-63e7-5b64-982d-0978163a7fe1", 00:30:57.865 "is_configured": true, 00:30:57.865 "data_offset": 0, 00:30:57.865 "data_size": 65536 00:30:57.865 }, 00:30:57.865 { 00:30:57.865 "name": "BaseBdev2", 00:30:57.865 "uuid": "803f16d6-fd3c-5841-9499-4fcf3fb23aa7", 00:30:57.865 "is_configured": true, 00:30:57.865 "data_offset": 0, 00:30:57.865 "data_size": 65536 00:30:57.865 } 00:30:57.865 ] 00:30:57.865 }' 00:30:57.865 13:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:57.865 13:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:30:57.865 13:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:58.123 13:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:30:58.123 13:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:30:58.123 13:51:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:58.123 13:51:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:30:58.123 [2024-11-20 13:51:04.938078] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:30:58.123 [2024-11-20 13:51:04.953648] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:30:58.123 [2024-11-20 13:51:05.055071] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:30:58.123 [2024-11-20 13:51:05.064636] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:58.123 [2024-11-20 13:51:05.064685] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:30:58.123 [2024-11-20 13:51:05.064698] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:30:58.123 [2024-11-20 13:51:05.105529] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006080 00:30:58.123 13:51:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:58.123 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:30:58.123 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:58.123 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:58.123 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:30:58.123 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:30:58.123 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:30:58.123 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:58.123 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:58.123 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:58.123 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:58.123 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:58.123 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:58.123 13:51:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:58.123 13:51:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:30:58.123 13:51:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:58.123 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:58.123 "name": "raid_bdev1", 00:30:58.123 "uuid": "a9479a2c-f98a-4794-8c37-b3989098e8da", 00:30:58.123 "strip_size_kb": 0, 00:30:58.123 "state": "online", 00:30:58.123 "raid_level": "raid1", 00:30:58.123 "superblock": false, 00:30:58.123 "num_base_bdevs": 2, 00:30:58.123 "num_base_bdevs_discovered": 1, 00:30:58.123 "num_base_bdevs_operational": 1, 00:30:58.123 "base_bdevs_list": [ 00:30:58.123 { 00:30:58.123 "name": null, 00:30:58.123 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:58.123 "is_configured": false, 00:30:58.123 "data_offset": 0, 00:30:58.123 "data_size": 65536 00:30:58.123 }, 00:30:58.123 { 00:30:58.123 "name": "BaseBdev2", 00:30:58.123 "uuid": "803f16d6-fd3c-5841-9499-4fcf3fb23aa7", 00:30:58.124 "is_configured": true, 00:30:58.124 "data_offset": 0, 00:30:58.124 "data_size": 65536 00:30:58.124 } 00:30:58.124 ] 00:30:58.124 }' 00:30:58.124 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:58.124 13:51:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:30:58.381 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:30:58.381 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:58.381 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:30:58.381 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:30:58.381 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:58.381 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:58.381 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:58.381 13:51:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:58.381 13:51:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:30:58.639 13:51:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:58.639 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:58.639 "name": "raid_bdev1", 00:30:58.639 "uuid": "a9479a2c-f98a-4794-8c37-b3989098e8da", 00:30:58.639 "strip_size_kb": 0, 00:30:58.639 "state": "online", 00:30:58.639 "raid_level": "raid1", 00:30:58.639 "superblock": false, 00:30:58.639 "num_base_bdevs": 2, 00:30:58.639 "num_base_bdevs_discovered": 1, 00:30:58.639 "num_base_bdevs_operational": 1, 00:30:58.639 "base_bdevs_list": [ 00:30:58.639 { 00:30:58.639 "name": null, 00:30:58.639 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:58.639 "is_configured": false, 00:30:58.639 "data_offset": 0, 00:30:58.639 "data_size": 65536 00:30:58.639 }, 00:30:58.639 { 00:30:58.639 "name": "BaseBdev2", 00:30:58.639 "uuid": "803f16d6-fd3c-5841-9499-4fcf3fb23aa7", 00:30:58.639 "is_configured": true, 00:30:58.639 "data_offset": 0, 00:30:58.639 "data_size": 65536 00:30:58.639 } 00:30:58.639 ] 00:30:58.639 }' 00:30:58.639 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:58.639 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:30:58.639 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:58.639 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:30:58.639 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:30:58.639 13:51:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:58.639 13:51:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:30:58.639 [2024-11-20 13:51:05.529771] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:30:58.639 13:51:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:58.639 13:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:30:58.639 [2024-11-20 13:51:05.594615] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:30:58.639 [2024-11-20 13:51:05.596697] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:30:58.897 154.50 IOPS, 463.50 MiB/s [2024-11-20T13:51:05.956Z] [2024-11-20 13:51:05.718343] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:30:58.897 [2024-11-20 13:51:05.718863] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:30:58.897 [2024-11-20 13:51:05.927692] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:30:58.897 [2024-11-20 13:51:05.928021] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:30:59.155 [2024-11-20 13:51:06.171387] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:30:59.155 [2024-11-20 13:51:06.171891] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:30:59.412 [2024-11-20 13:51:06.385797] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:30:59.712 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:30:59.712 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:59.712 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:30:59.712 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:30:59.712 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:59.712 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:59.712 13:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:59.712 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:59.712 13:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:30:59.712 144.33 IOPS, 433.00 MiB/s [2024-11-20T13:51:06.771Z] 13:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:59.712 [2024-11-20 13:51:06.607253] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:30:59.712 [2024-11-20 13:51:06.607740] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:30:59.712 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:59.712 "name": "raid_bdev1", 00:30:59.712 "uuid": "a9479a2c-f98a-4794-8c37-b3989098e8da", 00:30:59.712 "strip_size_kb": 0, 00:30:59.712 "state": "online", 00:30:59.712 "raid_level": "raid1", 00:30:59.712 "superblock": false, 00:30:59.712 "num_base_bdevs": 2, 00:30:59.712 "num_base_bdevs_discovered": 2, 00:30:59.712 "num_base_bdevs_operational": 2, 00:30:59.712 "process": { 00:30:59.712 "type": "rebuild", 00:30:59.712 "target": "spare", 00:30:59.712 "progress": { 00:30:59.712 "blocks": 12288, 00:30:59.712 "percent": 18 00:30:59.712 } 00:30:59.712 }, 00:30:59.712 "base_bdevs_list": [ 00:30:59.712 { 00:30:59.712 "name": "spare", 00:30:59.712 "uuid": "3dd90581-63e7-5b64-982d-0978163a7fe1", 00:30:59.712 "is_configured": true, 00:30:59.712 "data_offset": 0, 00:30:59.712 "data_size": 65536 00:30:59.712 }, 00:30:59.712 { 00:30:59.712 "name": "BaseBdev2", 00:30:59.712 "uuid": "803f16d6-fd3c-5841-9499-4fcf3fb23aa7", 00:30:59.712 "is_configured": true, 00:30:59.712 "data_offset": 0, 00:30:59.712 "data_size": 65536 00:30:59.712 } 00:30:59.712 ] 00:30:59.712 }' 00:30:59.712 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:59.712 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:30:59.712 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:59.712 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:30:59.712 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:30:59.712 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:30:59.712 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:30:59.712 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:30:59.712 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # local timeout=313 00:30:59.712 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:30:59.712 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:30:59.712 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:59.712 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:30:59.712 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:30:59.712 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:59.713 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:59.713 13:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:59.713 13:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:30:59.713 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:59.713 13:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:59.713 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:59.713 "name": "raid_bdev1", 00:30:59.713 "uuid": "a9479a2c-f98a-4794-8c37-b3989098e8da", 00:30:59.713 "strip_size_kb": 0, 00:30:59.713 "state": "online", 00:30:59.713 "raid_level": "raid1", 00:30:59.713 "superblock": false, 00:30:59.713 "num_base_bdevs": 2, 00:30:59.713 "num_base_bdevs_discovered": 2, 00:30:59.713 "num_base_bdevs_operational": 2, 00:30:59.713 "process": { 00:30:59.713 "type": "rebuild", 00:30:59.713 "target": "spare", 00:30:59.713 "progress": { 00:30:59.713 "blocks": 14336, 00:30:59.713 "percent": 21 00:30:59.713 } 00:30:59.713 }, 00:30:59.713 "base_bdevs_list": [ 00:30:59.713 { 00:30:59.713 "name": "spare", 00:30:59.713 "uuid": "3dd90581-63e7-5b64-982d-0978163a7fe1", 00:30:59.713 "is_configured": true, 00:30:59.713 "data_offset": 0, 00:30:59.713 "data_size": 65536 00:30:59.713 }, 00:30:59.713 { 00:30:59.713 "name": "BaseBdev2", 00:30:59.713 "uuid": "803f16d6-fd3c-5841-9499-4fcf3fb23aa7", 00:30:59.713 "is_configured": true, 00:30:59.713 "data_offset": 0, 00:30:59.713 "data_size": 65536 00:30:59.713 } 00:30:59.713 ] 00:30:59.713 }' 00:30:59.713 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:59.713 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:30:59.713 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:59.991 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:30:59.991 13:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:30:59.991 [2024-11-20 13:51:06.811960] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:31:00.248 [2024-11-20 13:51:07.235453] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:31:00.818 126.00 IOPS, 378.00 MiB/s [2024-11-20T13:51:07.877Z] 13:51:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:00.818 13:51:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:00.818 13:51:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:00.818 13:51:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:00.818 13:51:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:00.818 13:51:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:00.818 13:51:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:00.818 13:51:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:00.818 13:51:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:00.818 13:51:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:31:00.818 13:51:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:00.818 13:51:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:00.818 "name": "raid_bdev1", 00:31:00.818 "uuid": "a9479a2c-f98a-4794-8c37-b3989098e8da", 00:31:00.818 "strip_size_kb": 0, 00:31:00.818 "state": "online", 00:31:00.818 "raid_level": "raid1", 00:31:00.818 "superblock": false, 00:31:00.818 "num_base_bdevs": 2, 00:31:00.818 "num_base_bdevs_discovered": 2, 00:31:00.818 "num_base_bdevs_operational": 2, 00:31:00.818 "process": { 00:31:00.818 "type": "rebuild", 00:31:00.818 "target": "spare", 00:31:00.818 "progress": { 00:31:00.818 "blocks": 30720, 00:31:00.818 "percent": 46 00:31:00.818 } 00:31:00.818 }, 00:31:00.818 "base_bdevs_list": [ 00:31:00.818 { 00:31:00.818 "name": "spare", 00:31:00.818 "uuid": "3dd90581-63e7-5b64-982d-0978163a7fe1", 00:31:00.818 "is_configured": true, 00:31:00.818 "data_offset": 0, 00:31:00.818 "data_size": 65536 00:31:00.818 }, 00:31:00.818 { 00:31:00.818 "name": "BaseBdev2", 00:31:00.818 "uuid": "803f16d6-fd3c-5841-9499-4fcf3fb23aa7", 00:31:00.818 "is_configured": true, 00:31:00.818 "data_offset": 0, 00:31:00.818 "data_size": 65536 00:31:00.818 } 00:31:00.818 ] 00:31:00.818 }' 00:31:00.818 13:51:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:00.818 13:51:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:00.818 13:51:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:00.818 13:51:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:00.818 13:51:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:31:01.077 [2024-11-20 13:51:07.891324] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:31:01.077 [2024-11-20 13:51:07.891609] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:31:01.642 [2024-11-20 13:51:08.567710] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 45056 offset_begin: 43008 offset_end: 49152 00:31:01.900 111.00 IOPS, 333.00 MiB/s [2024-11-20T13:51:08.959Z] 13:51:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:01.900 13:51:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:01.900 13:51:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:01.900 13:51:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:01.900 13:51:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:01.900 13:51:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:01.900 13:51:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:01.900 13:51:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:01.900 13:51:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:01.900 13:51:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:31:01.900 13:51:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:01.900 13:51:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:01.900 "name": "raid_bdev1", 00:31:01.900 "uuid": "a9479a2c-f98a-4794-8c37-b3989098e8da", 00:31:01.900 "strip_size_kb": 0, 00:31:01.900 "state": "online", 00:31:01.900 "raid_level": "raid1", 00:31:01.900 "superblock": false, 00:31:01.900 "num_base_bdevs": 2, 00:31:01.900 "num_base_bdevs_discovered": 2, 00:31:01.900 "num_base_bdevs_operational": 2, 00:31:01.900 "process": { 00:31:01.900 "type": "rebuild", 00:31:01.900 "target": "spare", 00:31:01.900 "progress": { 00:31:01.900 "blocks": 47104, 00:31:01.900 "percent": 71 00:31:01.900 } 00:31:01.900 }, 00:31:01.900 "base_bdevs_list": [ 00:31:01.900 { 00:31:01.900 "name": "spare", 00:31:01.900 "uuid": "3dd90581-63e7-5b64-982d-0978163a7fe1", 00:31:01.900 "is_configured": true, 00:31:01.900 "data_offset": 0, 00:31:01.900 "data_size": 65536 00:31:01.900 }, 00:31:01.900 { 00:31:01.900 "name": "BaseBdev2", 00:31:01.900 "uuid": "803f16d6-fd3c-5841-9499-4fcf3fb23aa7", 00:31:01.900 "is_configured": true, 00:31:01.900 "data_offset": 0, 00:31:01.900 "data_size": 65536 00:31:01.900 } 00:31:01.900 ] 00:31:01.900 }' 00:31:01.900 13:51:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:01.900 13:51:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:01.900 13:51:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:01.900 13:51:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:01.900 13:51:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:31:02.157 [2024-11-20 13:51:09.127820] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:31:02.414 [2024-11-20 13:51:09.447139] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 57344 offset_begin: 55296 offset_end: 61440 00:31:02.672 99.17 IOPS, 297.50 MiB/s [2024-11-20T13:51:09.731Z] [2024-11-20 13:51:09.665303] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 59392 offset_begin: 55296 offset_end: 61440 00:31:02.930 13:51:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:02.930 13:51:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:02.931 13:51:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:02.931 13:51:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:02.931 13:51:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:02.931 13:51:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:02.931 13:51:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:02.931 13:51:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:02.931 13:51:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:31:02.931 13:51:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:02.931 13:51:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:02.931 13:51:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:02.931 "name": "raid_bdev1", 00:31:02.931 "uuid": "a9479a2c-f98a-4794-8c37-b3989098e8da", 00:31:02.931 "strip_size_kb": 0, 00:31:02.931 "state": "online", 00:31:02.931 "raid_level": "raid1", 00:31:02.931 "superblock": false, 00:31:02.931 "num_base_bdevs": 2, 00:31:02.931 "num_base_bdevs_discovered": 2, 00:31:02.931 "num_base_bdevs_operational": 2, 00:31:02.931 "process": { 00:31:02.931 "type": "rebuild", 00:31:02.931 "target": "spare", 00:31:02.931 "progress": { 00:31:02.931 "blocks": 61440, 00:31:02.931 "percent": 93 00:31:02.931 } 00:31:02.931 }, 00:31:02.931 "base_bdevs_list": [ 00:31:02.931 { 00:31:02.931 "name": "spare", 00:31:02.931 "uuid": "3dd90581-63e7-5b64-982d-0978163a7fe1", 00:31:02.931 "is_configured": true, 00:31:02.931 "data_offset": 0, 00:31:02.931 "data_size": 65536 00:31:02.931 }, 00:31:02.931 { 00:31:02.931 "name": "BaseBdev2", 00:31:02.931 "uuid": "803f16d6-fd3c-5841-9499-4fcf3fb23aa7", 00:31:02.931 "is_configured": true, 00:31:02.931 "data_offset": 0, 00:31:02.931 "data_size": 65536 00:31:02.931 } 00:31:02.931 ] 00:31:02.931 }' 00:31:02.931 13:51:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:03.261 13:51:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:03.261 13:51:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:03.261 13:51:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:03.261 13:51:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:31:03.261 [2024-11-20 13:51:10.097180] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:31:03.261 [2024-11-20 13:51:10.197205] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:31:03.261 [2024-11-20 13:51:10.204962] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:04.089 89.29 IOPS, 267.86 MiB/s [2024-11-20T13:51:11.148Z] 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:04.089 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:04.089 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:04.089 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:04.089 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:04.089 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:04.089 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:04.089 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:04.089 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:04.089 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:31:04.089 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:04.089 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:04.089 "name": "raid_bdev1", 00:31:04.089 "uuid": "a9479a2c-f98a-4794-8c37-b3989098e8da", 00:31:04.089 "strip_size_kb": 0, 00:31:04.089 "state": "online", 00:31:04.089 "raid_level": "raid1", 00:31:04.089 "superblock": false, 00:31:04.089 "num_base_bdevs": 2, 00:31:04.089 "num_base_bdevs_discovered": 2, 00:31:04.089 "num_base_bdevs_operational": 2, 00:31:04.089 "base_bdevs_list": [ 00:31:04.089 { 00:31:04.089 "name": "spare", 00:31:04.089 "uuid": "3dd90581-63e7-5b64-982d-0978163a7fe1", 00:31:04.089 "is_configured": true, 00:31:04.089 "data_offset": 0, 00:31:04.089 "data_size": 65536 00:31:04.089 }, 00:31:04.089 { 00:31:04.089 "name": "BaseBdev2", 00:31:04.089 "uuid": "803f16d6-fd3c-5841-9499-4fcf3fb23aa7", 00:31:04.089 "is_configured": true, 00:31:04.089 "data_offset": 0, 00:31:04.089 "data_size": 65536 00:31:04.089 } 00:31:04.089 ] 00:31:04.089 }' 00:31:04.089 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:04.089 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:31:04.089 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@709 -- # break 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:04.347 "name": "raid_bdev1", 00:31:04.347 "uuid": "a9479a2c-f98a-4794-8c37-b3989098e8da", 00:31:04.347 "strip_size_kb": 0, 00:31:04.347 "state": "online", 00:31:04.347 "raid_level": "raid1", 00:31:04.347 "superblock": false, 00:31:04.347 "num_base_bdevs": 2, 00:31:04.347 "num_base_bdevs_discovered": 2, 00:31:04.347 "num_base_bdevs_operational": 2, 00:31:04.347 "base_bdevs_list": [ 00:31:04.347 { 00:31:04.347 "name": "spare", 00:31:04.347 "uuid": "3dd90581-63e7-5b64-982d-0978163a7fe1", 00:31:04.347 "is_configured": true, 00:31:04.347 "data_offset": 0, 00:31:04.347 "data_size": 65536 00:31:04.347 }, 00:31:04.347 { 00:31:04.347 "name": "BaseBdev2", 00:31:04.347 "uuid": "803f16d6-fd3c-5841-9499-4fcf3fb23aa7", 00:31:04.347 "is_configured": true, 00:31:04.347 "data_offset": 0, 00:31:04.347 "data_size": 65536 00:31:04.347 } 00:31:04.347 ] 00:31:04.347 }' 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:04.347 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:04.347 "name": "raid_bdev1", 00:31:04.347 "uuid": "a9479a2c-f98a-4794-8c37-b3989098e8da", 00:31:04.347 "strip_size_kb": 0, 00:31:04.347 "state": "online", 00:31:04.347 "raid_level": "raid1", 00:31:04.347 "superblock": false, 00:31:04.347 "num_base_bdevs": 2, 00:31:04.347 "num_base_bdevs_discovered": 2, 00:31:04.348 "num_base_bdevs_operational": 2, 00:31:04.348 "base_bdevs_list": [ 00:31:04.348 { 00:31:04.348 "name": "spare", 00:31:04.348 "uuid": "3dd90581-63e7-5b64-982d-0978163a7fe1", 00:31:04.348 "is_configured": true, 00:31:04.348 "data_offset": 0, 00:31:04.348 "data_size": 65536 00:31:04.348 }, 00:31:04.348 { 00:31:04.348 "name": "BaseBdev2", 00:31:04.348 "uuid": "803f16d6-fd3c-5841-9499-4fcf3fb23aa7", 00:31:04.348 "is_configured": true, 00:31:04.348 "data_offset": 0, 00:31:04.348 "data_size": 65536 00:31:04.348 } 00:31:04.348 ] 00:31:04.348 }' 00:31:04.348 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:04.348 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:31:04.606 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:31:04.606 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:04.606 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:31:04.606 [2024-11-20 13:51:11.583944] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:31:04.606 [2024-11-20 13:51:11.584119] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:31:04.606 82.00 IOPS, 246.00 MiB/s 00:31:04.606 Latency(us) 00:31:04.606 [2024-11-20T13:51:11.665Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:31:04.606 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:31:04.606 raid_bdev1 : 8.02 82.09 246.26 0.00 0.00 17557.37 277.27 116149.96 00:31:04.606 [2024-11-20T13:51:11.665Z] =================================================================================================================== 00:31:04.606 [2024-11-20T13:51:11.665Z] Total : 82.09 246.26 0.00 0.00 17557.37 277.27 116149.96 00:31:04.606 [2024-11-20 13:51:11.608781] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:31:04.606 [2024-11-20 13:51:11.608933] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:04.606 [2024-11-20 13:51:11.609042] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:31:04.606 [2024-11-20 13:51:11.609198] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, sta{ 00:31:04.606 "results": [ 00:31:04.606 { 00:31:04.606 "job": "raid_bdev1", 00:31:04.606 "core_mask": "0x1", 00:31:04.606 "workload": "randrw", 00:31:04.606 "percentage": 50, 00:31:04.606 "status": "finished", 00:31:04.606 "queue_depth": 2, 00:31:04.606 "io_size": 3145728, 00:31:04.606 "runtime": 8.015921, 00:31:04.606 "iops": 82.08663733088189, 00:31:04.606 "mibps": 246.25991199264564, 00:31:04.606 "io_failed": 0, 00:31:04.606 "io_timeout": 0, 00:31:04.606 "avg_latency_us": 17557.37184007482, 00:31:04.606 "min_latency_us": 277.2676923076923, 00:31:04.606 "max_latency_us": 116149.95692307693 00:31:04.606 } 00:31:04.606 ], 00:31:04.606 "core_count": 1 00:31:04.606 } 00:31:04.606 te offline 00:31:04.606 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:04.606 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:04.606 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # jq length 00:31:04.606 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:04.606 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:31:04.606 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:04.606 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:31:04.606 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:31:04.606 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:31:04.606 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:31:04.606 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:31:04.606 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:31:04.606 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:31:04.606 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:31:04.606 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:31:04.606 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:31:04.606 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:31:04.606 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:31:04.606 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:31:04.865 /dev/nbd0 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:31:04.865 1+0 records in 00:31:04.865 1+0 records out 00:31:04.865 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000316992 s, 12.9 MB/s 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev2 ']' 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev2 /dev/nbd1 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev2') 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:31:04.865 13:51:11 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev2 /dev/nbd1 00:31:05.124 /dev/nbd1 00:31:05.124 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:31:05.124 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:31:05.124 13:51:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:31:05.124 13:51:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:31:05.124 13:51:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:31:05.124 13:51:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:31:05.124 13:51:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:31:05.124 13:51:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:31:05.124 13:51:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:31:05.124 13:51:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:31:05.124 13:51:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:31:05.124 1+0 records in 00:31:05.124 1+0 records out 00:31:05.124 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000313785 s, 13.1 MB/s 00:31:05.124 13:51:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:05.124 13:51:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:31:05.124 13:51:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:05.124 13:51:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:31:05.124 13:51:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:31:05.124 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:31:05.124 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:31:05.124 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:31:05.382 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:31:05.382 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:31:05.382 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:31:05.382 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:31:05.382 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:31:05.382 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:31:05.382 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@784 -- # killprocess 74319 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@954 -- # '[' -z 74319 ']' 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@958 -- # kill -0 74319 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # uname 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 74319 00:31:05.641 killing process with pid 74319 00:31:05.641 Received shutdown signal, test time was about 9.116015 seconds 00:31:05.641 00:31:05.641 Latency(us) 00:31:05.641 [2024-11-20T13:51:12.700Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:31:05.641 [2024-11-20T13:51:12.700Z] =================================================================================================================== 00:31:05.641 [2024-11-20T13:51:12.700Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:31:05.641 13:51:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:31:05.642 13:51:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:31:05.642 13:51:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 74319' 00:31:05.642 13:51:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@973 -- # kill 74319 00:31:05.642 [2024-11-20 13:51:12.695991] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:31:05.642 13:51:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@978 -- # wait 74319 00:31:05.901 [2024-11-20 13:51:12.810941] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:31:06.468 13:51:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@786 -- # return 0 00:31:06.468 00:31:06.468 real 0m11.378s 00:31:06.468 user 0m13.830s 00:31:06.468 sys 0m1.083s 00:31:06.468 13:51:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:31:06.468 13:51:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:31:06.468 ************************************ 00:31:06.468 END TEST raid_rebuild_test_io 00:31:06.468 ************************************ 00:31:06.468 13:51:13 bdev_raid -- bdev/bdev_raid.sh@981 -- # run_test raid_rebuild_test_sb_io raid_rebuild_test raid1 2 true true true 00:31:06.468 13:51:13 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:31:06.468 13:51:13 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:31:06.468 13:51:13 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:31:06.468 ************************************ 00:31:06.468 START TEST raid_rebuild_test_sb_io 00:31:06.468 ************************************ 00:31:06.468 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true true true 00:31:06.468 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:31:06.468 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:31:06.468 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:31:06.468 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:31:06.468 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:31:06.468 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:31:06.468 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:31:06.468 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:31:06.468 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:31:06.468 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:31:06.468 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:31:06.468 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:31:06.468 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:31:06.468 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:31:06.468 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:31:06.468 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:31:06.469 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:31:06.469 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:31:06.469 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:31:06.469 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:31:06.469 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:31:06.469 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:31:06.469 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:31:06.469 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:31:06.469 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:31:06.469 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@597 -- # raid_pid=74699 00:31:06.469 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 74699 00:31:06.469 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@835 -- # '[' -z 74699 ']' 00:31:06.469 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:31:06.469 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:31:06.469 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:31:06.469 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:31:06.469 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:06.469 13:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:31:06.739 I/O size of 3145728 is greater than zero copy threshold (65536). 00:31:06.739 Zero copy mechanism will not be used. 00:31:06.739 [2024-11-20 13:51:13.564449] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:31:06.739 [2024-11-20 13:51:13.564579] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74699 ] 00:31:06.739 [2024-11-20 13:51:13.721511] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:31:06.997 [2024-11-20 13:51:13.821326] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:31:06.997 [2024-11-20 13:51:13.941498] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:31:06.997 [2024-11-20 13:51:13.941561] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:31:07.562 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:31:07.562 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # return 0 00:31:07.562 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:31:07.562 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:31:07.562 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:07.562 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:07.562 BaseBdev1_malloc 00:31:07.562 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:07.562 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:31:07.562 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:07.562 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:07.562 [2024-11-20 13:51:14.435049] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:31:07.562 [2024-11-20 13:51:14.435262] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:07.563 [2024-11-20 13:51:14.435289] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:31:07.563 [2024-11-20 13:51:14.435300] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:07.563 [2024-11-20 13:51:14.437202] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:07.563 [2024-11-20 13:51:14.437233] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:31:07.563 BaseBdev1 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:07.563 BaseBdev2_malloc 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:07.563 [2024-11-20 13:51:14.468675] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:31:07.563 [2024-11-20 13:51:14.468729] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:07.563 [2024-11-20 13:51:14.468747] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:31:07.563 [2024-11-20 13:51:14.468757] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:07.563 [2024-11-20 13:51:14.470644] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:07.563 [2024-11-20 13:51:14.470805] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:31:07.563 BaseBdev2 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:07.563 spare_malloc 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:07.563 spare_delay 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:07.563 [2024-11-20 13:51:14.528063] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:31:07.563 [2024-11-20 13:51:14.528275] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:07.563 [2024-11-20 13:51:14.528298] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:31:07.563 [2024-11-20 13:51:14.528308] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:07.563 [2024-11-20 13:51:14.530296] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:07.563 [2024-11-20 13:51:14.530330] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:31:07.563 spare 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:07.563 [2024-11-20 13:51:14.536111] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:31:07.563 [2024-11-20 13:51:14.537762] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:31:07.563 [2024-11-20 13:51:14.537907] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:31:07.563 [2024-11-20 13:51:14.537919] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:31:07.563 [2024-11-20 13:51:14.538157] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:31:07.563 [2024-11-20 13:51:14.538301] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:31:07.563 [2024-11-20 13:51:14.538309] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:31:07.563 [2024-11-20 13:51:14.538434] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:07.563 "name": "raid_bdev1", 00:31:07.563 "uuid": "5100785b-7326-4f0a-9446-6e93c7dce5da", 00:31:07.563 "strip_size_kb": 0, 00:31:07.563 "state": "online", 00:31:07.563 "raid_level": "raid1", 00:31:07.563 "superblock": true, 00:31:07.563 "num_base_bdevs": 2, 00:31:07.563 "num_base_bdevs_discovered": 2, 00:31:07.563 "num_base_bdevs_operational": 2, 00:31:07.563 "base_bdevs_list": [ 00:31:07.563 { 00:31:07.563 "name": "BaseBdev1", 00:31:07.563 "uuid": "45e28997-b36b-593d-9ca3-97152f061e69", 00:31:07.563 "is_configured": true, 00:31:07.563 "data_offset": 2048, 00:31:07.563 "data_size": 63488 00:31:07.563 }, 00:31:07.563 { 00:31:07.563 "name": "BaseBdev2", 00:31:07.563 "uuid": "80ea6115-d09e-573d-8300-abe8f8bd1084", 00:31:07.563 "is_configured": true, 00:31:07.563 "data_offset": 2048, 00:31:07.563 "data_size": 63488 00:31:07.563 } 00:31:07.563 ] 00:31:07.563 }' 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:07.563 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:07.821 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:31:07.821 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:31:07.821 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:07.821 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:07.821 [2024-11-20 13:51:14.856451] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:31:07.821 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:08.079 [2024-11-20 13:51:14.920184] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:08.079 "name": "raid_bdev1", 00:31:08.079 "uuid": "5100785b-7326-4f0a-9446-6e93c7dce5da", 00:31:08.079 "strip_size_kb": 0, 00:31:08.079 "state": "online", 00:31:08.079 "raid_level": "raid1", 00:31:08.079 "superblock": true, 00:31:08.079 "num_base_bdevs": 2, 00:31:08.079 "num_base_bdevs_discovered": 1, 00:31:08.079 "num_base_bdevs_operational": 1, 00:31:08.079 "base_bdevs_list": [ 00:31:08.079 { 00:31:08.079 "name": null, 00:31:08.079 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:08.079 "is_configured": false, 00:31:08.079 "data_offset": 0, 00:31:08.079 "data_size": 63488 00:31:08.079 }, 00:31:08.079 { 00:31:08.079 "name": "BaseBdev2", 00:31:08.079 "uuid": "80ea6115-d09e-573d-8300-abe8f8bd1084", 00:31:08.079 "is_configured": true, 00:31:08.079 "data_offset": 2048, 00:31:08.079 "data_size": 63488 00:31:08.079 } 00:31:08.079 ] 00:31:08.079 }' 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:08.079 13:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:08.079 I/O size of 3145728 is greater than zero copy threshold (65536). 00:31:08.079 Zero copy mechanism will not be used. 00:31:08.079 Running I/O for 60 seconds... 00:31:08.079 [2024-11-20 13:51:15.049215] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:31:08.338 13:51:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:31:08.338 13:51:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:08.338 13:51:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:08.338 [2024-11-20 13:51:15.251407] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:31:08.338 13:51:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:08.338 13:51:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:31:08.338 [2024-11-20 13:51:15.302815] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:31:08.338 [2024-11-20 13:51:15.304762] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:31:08.598 [2024-11-20 13:51:15.416914] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:31:08.598 [2024-11-20 13:51:15.417645] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:31:08.598 [2024-11-20 13:51:15.527708] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:31:08.598 [2024-11-20 13:51:15.528139] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:31:08.857 [2024-11-20 13:51:15.861038] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:31:09.374 137.00 IOPS, 411.00 MiB/s [2024-11-20T13:51:16.433Z] 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:09.374 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:09.374 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:09.374 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:09.374 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:09.374 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:09.374 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:09.374 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:09.374 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:09.374 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:09.374 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:09.374 "name": "raid_bdev1", 00:31:09.374 "uuid": "5100785b-7326-4f0a-9446-6e93c7dce5da", 00:31:09.374 "strip_size_kb": 0, 00:31:09.374 "state": "online", 00:31:09.374 "raid_level": "raid1", 00:31:09.374 "superblock": true, 00:31:09.374 "num_base_bdevs": 2, 00:31:09.374 "num_base_bdevs_discovered": 2, 00:31:09.374 "num_base_bdevs_operational": 2, 00:31:09.374 "process": { 00:31:09.374 "type": "rebuild", 00:31:09.374 "target": "spare", 00:31:09.374 "progress": { 00:31:09.374 "blocks": 12288, 00:31:09.374 "percent": 19 00:31:09.374 } 00:31:09.374 }, 00:31:09.374 "base_bdevs_list": [ 00:31:09.374 { 00:31:09.374 "name": "spare", 00:31:09.374 "uuid": "ce0a4bcc-0059-54cb-bd4a-3455b477408c", 00:31:09.374 "is_configured": true, 00:31:09.374 "data_offset": 2048, 00:31:09.374 "data_size": 63488 00:31:09.374 }, 00:31:09.374 { 00:31:09.374 "name": "BaseBdev2", 00:31:09.374 "uuid": "80ea6115-d09e-573d-8300-abe8f8bd1084", 00:31:09.374 "is_configured": true, 00:31:09.374 "data_offset": 2048, 00:31:09.374 "data_size": 63488 00:31:09.374 } 00:31:09.374 ] 00:31:09.374 }' 00:31:09.374 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:09.374 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:09.374 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:09.374 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:09.374 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:31:09.374 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:09.374 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:09.374 [2024-11-20 13:51:16.393001] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:31:09.632 [2024-11-20 13:51:16.540029] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:31:09.632 [2024-11-20 13:51:16.542869] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:09.632 [2024-11-20 13:51:16.543002] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:31:09.632 [2024-11-20 13:51:16.543030] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:31:09.632 [2024-11-20 13:51:16.566045] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006080 00:31:09.633 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:09.633 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:31:09.633 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:09.633 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:09.633 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:09.633 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:09.633 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:31:09.633 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:09.633 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:09.633 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:09.633 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:09.633 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:09.633 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:09.633 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:09.633 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:09.633 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:09.633 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:09.633 "name": "raid_bdev1", 00:31:09.633 "uuid": "5100785b-7326-4f0a-9446-6e93c7dce5da", 00:31:09.633 "strip_size_kb": 0, 00:31:09.633 "state": "online", 00:31:09.633 "raid_level": "raid1", 00:31:09.633 "superblock": true, 00:31:09.633 "num_base_bdevs": 2, 00:31:09.633 "num_base_bdevs_discovered": 1, 00:31:09.633 "num_base_bdevs_operational": 1, 00:31:09.633 "base_bdevs_list": [ 00:31:09.633 { 00:31:09.633 "name": null, 00:31:09.633 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:09.633 "is_configured": false, 00:31:09.633 "data_offset": 0, 00:31:09.633 "data_size": 63488 00:31:09.633 }, 00:31:09.633 { 00:31:09.633 "name": "BaseBdev2", 00:31:09.633 "uuid": "80ea6115-d09e-573d-8300-abe8f8bd1084", 00:31:09.633 "is_configured": true, 00:31:09.633 "data_offset": 2048, 00:31:09.633 "data_size": 63488 00:31:09.633 } 00:31:09.633 ] 00:31:09.633 }' 00:31:09.633 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:09.633 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:09.891 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:31:09.891 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:09.891 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:31:09.891 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:31:09.891 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:09.891 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:09.891 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:09.891 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:09.891 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:09.891 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:09.891 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:09.891 "name": "raid_bdev1", 00:31:09.891 "uuid": "5100785b-7326-4f0a-9446-6e93c7dce5da", 00:31:09.891 "strip_size_kb": 0, 00:31:09.891 "state": "online", 00:31:09.891 "raid_level": "raid1", 00:31:09.891 "superblock": true, 00:31:09.891 "num_base_bdevs": 2, 00:31:09.891 "num_base_bdevs_discovered": 1, 00:31:09.891 "num_base_bdevs_operational": 1, 00:31:09.891 "base_bdevs_list": [ 00:31:09.891 { 00:31:09.891 "name": null, 00:31:09.891 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:09.891 "is_configured": false, 00:31:09.891 "data_offset": 0, 00:31:09.891 "data_size": 63488 00:31:09.891 }, 00:31:09.891 { 00:31:09.891 "name": "BaseBdev2", 00:31:09.891 "uuid": "80ea6115-d09e-573d-8300-abe8f8bd1084", 00:31:09.891 "is_configured": true, 00:31:09.891 "data_offset": 2048, 00:31:09.891 "data_size": 63488 00:31:09.891 } 00:31:09.891 ] 00:31:09.891 }' 00:31:09.891 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:09.891 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:31:09.891 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:10.150 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:31:10.150 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:31:10.150 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:10.150 13:51:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:10.150 [2024-11-20 13:51:16.980881] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:31:10.150 13:51:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:10.150 13:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:31:10.150 [2024-11-20 13:51:17.010704] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:31:10.150 [2024-11-20 13:51:17.012422] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:31:10.150 169.50 IOPS, 508.50 MiB/s [2024-11-20T13:51:17.209Z] [2024-11-20 13:51:17.118188] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:31:10.150 [2024-11-20 13:51:17.118659] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:31:10.408 [2024-11-20 13:51:17.325999] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:31:10.408 [2024-11-20 13:51:17.326464] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:31:10.665 [2024-11-20 13:51:17.654140] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:31:10.922 [2024-11-20 13:51:17.755394] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:31:10.922 [2024-11-20 13:51:17.755723] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:31:11.180 [2024-11-20 13:51:17.992299] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:31:11.180 [2024-11-20 13:51:17.992795] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:11.180 "name": "raid_bdev1", 00:31:11.180 "uuid": "5100785b-7326-4f0a-9446-6e93c7dce5da", 00:31:11.180 "strip_size_kb": 0, 00:31:11.180 "state": "online", 00:31:11.180 "raid_level": "raid1", 00:31:11.180 "superblock": true, 00:31:11.180 "num_base_bdevs": 2, 00:31:11.180 "num_base_bdevs_discovered": 2, 00:31:11.180 "num_base_bdevs_operational": 2, 00:31:11.180 "process": { 00:31:11.180 "type": "rebuild", 00:31:11.180 "target": "spare", 00:31:11.180 "progress": { 00:31:11.180 "blocks": 14336, 00:31:11.180 "percent": 22 00:31:11.180 } 00:31:11.180 }, 00:31:11.180 "base_bdevs_list": [ 00:31:11.180 { 00:31:11.180 "name": "spare", 00:31:11.180 "uuid": "ce0a4bcc-0059-54cb-bd4a-3455b477408c", 00:31:11.180 "is_configured": true, 00:31:11.180 "data_offset": 2048, 00:31:11.180 "data_size": 63488 00:31:11.180 }, 00:31:11.180 { 00:31:11.180 "name": "BaseBdev2", 00:31:11.180 "uuid": "80ea6115-d09e-573d-8300-abe8f8bd1084", 00:31:11.180 "is_configured": true, 00:31:11.180 "data_offset": 2048, 00:31:11.180 "data_size": 63488 00:31:11.180 } 00:31:11.180 ] 00:31:11.180 }' 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:11.180 141.33 IOPS, 424.00 MiB/s [2024-11-20T13:51:18.239Z] 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:31:11.180 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # local timeout=325 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:11.180 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:11.180 "name": "raid_bdev1", 00:31:11.181 "uuid": "5100785b-7326-4f0a-9446-6e93c7dce5da", 00:31:11.181 "strip_size_kb": 0, 00:31:11.181 "state": "online", 00:31:11.181 "raid_level": "raid1", 00:31:11.181 "superblock": true, 00:31:11.181 "num_base_bdevs": 2, 00:31:11.181 "num_base_bdevs_discovered": 2, 00:31:11.181 "num_base_bdevs_operational": 2, 00:31:11.181 "process": { 00:31:11.181 "type": "rebuild", 00:31:11.181 "target": "spare", 00:31:11.181 "progress": { 00:31:11.181 "blocks": 14336, 00:31:11.181 "percent": 22 00:31:11.181 } 00:31:11.181 }, 00:31:11.181 "base_bdevs_list": [ 00:31:11.181 { 00:31:11.181 "name": "spare", 00:31:11.181 "uuid": "ce0a4bcc-0059-54cb-bd4a-3455b477408c", 00:31:11.181 "is_configured": true, 00:31:11.181 "data_offset": 2048, 00:31:11.181 "data_size": 63488 00:31:11.181 }, 00:31:11.181 { 00:31:11.181 "name": "BaseBdev2", 00:31:11.181 "uuid": "80ea6115-d09e-573d-8300-abe8f8bd1084", 00:31:11.181 "is_configured": true, 00:31:11.181 "data_offset": 2048, 00:31:11.181 "data_size": 63488 00:31:11.181 } 00:31:11.181 ] 00:31:11.181 }' 00:31:11.181 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:11.181 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:11.181 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:11.181 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:11.181 13:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:31:11.181 [2024-11-20 13:51:18.213238] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:31:11.438 [2024-11-20 13:51:18.456322] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:31:12.004 [2024-11-20 13:51:18.778547] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:31:12.262 127.25 IOPS, 381.75 MiB/s [2024-11-20T13:51:19.321Z] 13:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:12.262 13:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:12.262 13:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:12.262 13:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:12.262 13:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:12.262 13:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:12.262 [2024-11-20 13:51:19.214024] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:31:12.262 13:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:12.262 13:51:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:12.262 13:51:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:12.262 13:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:12.262 13:51:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:12.262 13:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:12.262 "name": "raid_bdev1", 00:31:12.262 "uuid": "5100785b-7326-4f0a-9446-6e93c7dce5da", 00:31:12.262 "strip_size_kb": 0, 00:31:12.262 "state": "online", 00:31:12.262 "raid_level": "raid1", 00:31:12.262 "superblock": true, 00:31:12.262 "num_base_bdevs": 2, 00:31:12.262 "num_base_bdevs_discovered": 2, 00:31:12.262 "num_base_bdevs_operational": 2, 00:31:12.262 "process": { 00:31:12.262 "type": "rebuild", 00:31:12.262 "target": "spare", 00:31:12.262 "progress": { 00:31:12.262 "blocks": 32768, 00:31:12.262 "percent": 51 00:31:12.262 } 00:31:12.262 }, 00:31:12.262 "base_bdevs_list": [ 00:31:12.262 { 00:31:12.262 "name": "spare", 00:31:12.262 "uuid": "ce0a4bcc-0059-54cb-bd4a-3455b477408c", 00:31:12.262 "is_configured": true, 00:31:12.262 "data_offset": 2048, 00:31:12.262 "data_size": 63488 00:31:12.262 }, 00:31:12.262 { 00:31:12.262 "name": "BaseBdev2", 00:31:12.262 "uuid": "80ea6115-d09e-573d-8300-abe8f8bd1084", 00:31:12.262 "is_configured": true, 00:31:12.262 "data_offset": 2048, 00:31:12.262 "data_size": 63488 00:31:12.262 } 00:31:12.262 ] 00:31:12.262 }' 00:31:12.262 13:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:12.262 13:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:12.262 13:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:12.262 13:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:12.262 13:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:31:13.197 109.00 IOPS, 327.00 MiB/s [2024-11-20T13:51:20.256Z] [2024-11-20 13:51:20.109668] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:31:13.455 13:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:13.455 13:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:13.455 13:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:13.455 13:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:13.455 13:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:13.455 13:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:13.455 13:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:13.455 13:51:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:13.455 13:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:13.455 13:51:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:13.455 13:51:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:13.455 13:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:13.455 "name": "raid_bdev1", 00:31:13.455 "uuid": "5100785b-7326-4f0a-9446-6e93c7dce5da", 00:31:13.455 "strip_size_kb": 0, 00:31:13.455 "state": "online", 00:31:13.455 "raid_level": "raid1", 00:31:13.455 "superblock": true, 00:31:13.455 "num_base_bdevs": 2, 00:31:13.455 "num_base_bdevs_discovered": 2, 00:31:13.455 "num_base_bdevs_operational": 2, 00:31:13.455 "process": { 00:31:13.455 "type": "rebuild", 00:31:13.455 "target": "spare", 00:31:13.455 "progress": { 00:31:13.455 "blocks": 47104, 00:31:13.455 "percent": 74 00:31:13.455 } 00:31:13.455 }, 00:31:13.455 "base_bdevs_list": [ 00:31:13.455 { 00:31:13.455 "name": "spare", 00:31:13.455 "uuid": "ce0a4bcc-0059-54cb-bd4a-3455b477408c", 00:31:13.455 "is_configured": true, 00:31:13.455 "data_offset": 2048, 00:31:13.455 "data_size": 63488 00:31:13.455 }, 00:31:13.455 { 00:31:13.455 "name": "BaseBdev2", 00:31:13.455 "uuid": "80ea6115-d09e-573d-8300-abe8f8bd1084", 00:31:13.455 "is_configured": true, 00:31:13.455 "data_offset": 2048, 00:31:13.455 "data_size": 63488 00:31:13.455 } 00:31:13.456 ] 00:31:13.456 }' 00:31:13.456 13:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:13.456 13:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:13.456 13:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:13.456 13:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:13.456 13:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:31:13.713 [2024-11-20 13:51:20.537036] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:31:13.971 [2024-11-20 13:51:20.859156] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 59392 offset_begin: 55296 offset_end: 61440 00:31:14.230 96.83 IOPS, 290.50 MiB/s [2024-11-20T13:51:21.289Z] [2024-11-20 13:51:21.079944] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:31:14.230 [2024-11-20 13:51:21.185170] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:31:14.230 [2024-11-20 13:51:21.187665] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:14.488 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:14.488 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:14.488 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:14.488 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:14.488 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:14.488 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:14.488 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:14.488 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:14.488 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:14.488 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:14.488 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:14.488 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:14.488 "name": "raid_bdev1", 00:31:14.488 "uuid": "5100785b-7326-4f0a-9446-6e93c7dce5da", 00:31:14.488 "strip_size_kb": 0, 00:31:14.488 "state": "online", 00:31:14.488 "raid_level": "raid1", 00:31:14.488 "superblock": true, 00:31:14.488 "num_base_bdevs": 2, 00:31:14.488 "num_base_bdevs_discovered": 2, 00:31:14.488 "num_base_bdevs_operational": 2, 00:31:14.488 "base_bdevs_list": [ 00:31:14.488 { 00:31:14.488 "name": "spare", 00:31:14.488 "uuid": "ce0a4bcc-0059-54cb-bd4a-3455b477408c", 00:31:14.488 "is_configured": true, 00:31:14.488 "data_offset": 2048, 00:31:14.488 "data_size": 63488 00:31:14.488 }, 00:31:14.488 { 00:31:14.488 "name": "BaseBdev2", 00:31:14.488 "uuid": "80ea6115-d09e-573d-8300-abe8f8bd1084", 00:31:14.488 "is_configured": true, 00:31:14.488 "data_offset": 2048, 00:31:14.488 "data_size": 63488 00:31:14.488 } 00:31:14.488 ] 00:31:14.488 }' 00:31:14.488 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:14.488 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:31:14.488 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:14.488 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:31:14.488 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@709 -- # break 00:31:14.489 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:31:14.489 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:14.489 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:31:14.489 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:31:14.489 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:14.489 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:14.489 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:14.489 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:14.489 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:14.489 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:14.489 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:14.489 "name": "raid_bdev1", 00:31:14.489 "uuid": "5100785b-7326-4f0a-9446-6e93c7dce5da", 00:31:14.489 "strip_size_kb": 0, 00:31:14.489 "state": "online", 00:31:14.489 "raid_level": "raid1", 00:31:14.489 "superblock": true, 00:31:14.489 "num_base_bdevs": 2, 00:31:14.489 "num_base_bdevs_discovered": 2, 00:31:14.489 "num_base_bdevs_operational": 2, 00:31:14.489 "base_bdevs_list": [ 00:31:14.489 { 00:31:14.489 "name": "spare", 00:31:14.489 "uuid": "ce0a4bcc-0059-54cb-bd4a-3455b477408c", 00:31:14.489 "is_configured": true, 00:31:14.489 "data_offset": 2048, 00:31:14.489 "data_size": 63488 00:31:14.489 }, 00:31:14.489 { 00:31:14.489 "name": "BaseBdev2", 00:31:14.489 "uuid": "80ea6115-d09e-573d-8300-abe8f8bd1084", 00:31:14.489 "is_configured": true, 00:31:14.489 "data_offset": 2048, 00:31:14.489 "data_size": 63488 00:31:14.489 } 00:31:14.489 ] 00:31:14.489 }' 00:31:14.746 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:14.746 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:31:14.746 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:14.746 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:31:14.746 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:31:14.746 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:14.746 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:14.746 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:14.746 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:14.746 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:31:14.746 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:14.747 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:14.747 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:14.747 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:14.747 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:14.747 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:14.747 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:14.747 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:14.747 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:14.747 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:14.747 "name": "raid_bdev1", 00:31:14.747 "uuid": "5100785b-7326-4f0a-9446-6e93c7dce5da", 00:31:14.747 "strip_size_kb": 0, 00:31:14.747 "state": "online", 00:31:14.747 "raid_level": "raid1", 00:31:14.747 "superblock": true, 00:31:14.747 "num_base_bdevs": 2, 00:31:14.747 "num_base_bdevs_discovered": 2, 00:31:14.747 "num_base_bdevs_operational": 2, 00:31:14.747 "base_bdevs_list": [ 00:31:14.747 { 00:31:14.747 "name": "spare", 00:31:14.747 "uuid": "ce0a4bcc-0059-54cb-bd4a-3455b477408c", 00:31:14.747 "is_configured": true, 00:31:14.747 "data_offset": 2048, 00:31:14.747 "data_size": 63488 00:31:14.747 }, 00:31:14.747 { 00:31:14.747 "name": "BaseBdev2", 00:31:14.747 "uuid": "80ea6115-d09e-573d-8300-abe8f8bd1084", 00:31:14.747 "is_configured": true, 00:31:14.747 "data_offset": 2048, 00:31:14.747 "data_size": 63488 00:31:14.747 } 00:31:14.747 ] 00:31:14.747 }' 00:31:14.747 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:14.747 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:15.004 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:31:15.004 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:15.004 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:15.004 [2024-11-20 13:51:21.916581] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:31:15.004 [2024-11-20 13:51:21.916614] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:31:15.004 00:31:15.004 Latency(us) 00:31:15.005 [2024-11-20T13:51:22.064Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:31:15.005 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:31:15.005 raid_bdev1 : 6.89 88.74 266.21 0.00 0.00 16065.36 239.46 113730.17 00:31:15.005 [2024-11-20T13:51:22.064Z] =================================================================================================================== 00:31:15.005 [2024-11-20T13:51:22.064Z] Total : 88.74 266.21 0.00 0.00 16065.36 239.46 113730.17 00:31:15.005 { 00:31:15.005 "results": [ 00:31:15.005 { 00:31:15.005 "job": "raid_bdev1", 00:31:15.005 "core_mask": "0x1", 00:31:15.005 "workload": "randrw", 00:31:15.005 "percentage": 50, 00:31:15.005 "status": "finished", 00:31:15.005 "queue_depth": 2, 00:31:15.005 "io_size": 3145728, 00:31:15.005 "runtime": 6.885422, 00:31:15.005 "iops": 88.73820660520154, 00:31:15.005 "mibps": 266.21461981560464, 00:31:15.005 "io_failed": 0, 00:31:15.005 "io_timeout": 0, 00:31:15.005 "avg_latency_us": 16065.359768349492, 00:31:15.005 "min_latency_us": 239.45846153846153, 00:31:15.005 "max_latency_us": 113730.16615384615 00:31:15.005 } 00:31:15.005 ], 00:31:15.005 "core_count": 1 00:31:15.005 } 00:31:15.005 [2024-11-20 13:51:21.949180] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:31:15.005 [2024-11-20 13:51:21.949250] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:15.005 [2024-11-20 13:51:21.949316] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:31:15.005 [2024-11-20 13:51:21.949329] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:31:15.005 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:15.005 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:15.005 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # jq length 00:31:15.005 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:15.005 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:15.005 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:15.005 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:31:15.005 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:31:15.005 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:31:15.005 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:31:15.005 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:31:15.005 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:31:15.005 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:31:15.005 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:31:15.005 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:31:15.005 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:31:15.005 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:31:15.005 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:31:15.005 13:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:31:15.263 /dev/nbd0 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:31:15.263 1+0 records in 00:31:15.263 1+0 records out 00:31:15.263 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000287125 s, 14.3 MB/s 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev2 ']' 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev2 /dev/nbd1 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev2') 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:31:15.263 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev2 /dev/nbd1 00:31:15.521 /dev/nbd1 00:31:15.521 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:31:15.521 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:31:15.521 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:31:15.521 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:31:15.521 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:31:15.521 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:31:15.521 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:31:15.521 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:31:15.521 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:31:15.521 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:31:15.521 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:31:15.521 1+0 records in 00:31:15.521 1+0 records out 00:31:15.521 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00034349 s, 11.9 MB/s 00:31:15.521 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:15.521 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:31:15.521 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:15.521 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:31:15.521 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:31:15.521 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:31:15.521 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:31:15.521 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:31:15.521 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:31:15.522 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:31:15.522 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:31:15.522 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:31:15.522 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:31:15.522 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:31:15.522 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:31:15.779 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:31:15.779 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:31:15.779 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:31:15.779 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:31:15.779 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:31:15.779 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:31:15.779 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:31:15.779 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:31:15.779 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:31:15.779 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:31:15.779 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:31:15.779 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:31:15.779 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:31:15.779 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:31:15.779 13:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:31:16.037 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:31:16.037 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:31:16.037 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:31:16.037 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:31:16.037 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:31:16.037 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:31:16.037 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:31:16.037 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:31:16.037 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:31:16.037 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:31:16.037 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:16.037 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:16.037 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:16.037 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:31:16.037 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:16.037 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:16.037 [2024-11-20 13:51:23.057748] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:31:16.037 [2024-11-20 13:51:23.057805] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:16.037 [2024-11-20 13:51:23.057827] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:31:16.037 [2024-11-20 13:51:23.057837] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:16.037 [2024-11-20 13:51:23.059839] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:16.037 [2024-11-20 13:51:23.059876] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:31:16.037 [2024-11-20 13:51:23.059958] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:31:16.037 [2024-11-20 13:51:23.060014] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:31:16.037 [2024-11-20 13:51:23.060126] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:31:16.037 spare 00:31:16.037 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:16.037 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:31:16.037 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:16.037 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:16.294 [2024-11-20 13:51:23.160223] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:31:16.294 [2024-11-20 13:51:23.160262] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:31:16.294 [2024-11-20 13:51:23.160573] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b0d0 00:31:16.294 [2024-11-20 13:51:23.160747] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:31:16.294 [2024-11-20 13:51:23.160758] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:31:16.294 [2024-11-20 13:51:23.160918] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:16.294 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:16.294 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:31:16.294 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:16.294 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:16.294 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:16.294 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:16.294 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:31:16.294 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:16.294 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:16.294 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:16.294 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:16.294 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:16.294 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:16.294 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:16.294 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:16.294 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:16.294 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:16.294 "name": "raid_bdev1", 00:31:16.294 "uuid": "5100785b-7326-4f0a-9446-6e93c7dce5da", 00:31:16.294 "strip_size_kb": 0, 00:31:16.294 "state": "online", 00:31:16.294 "raid_level": "raid1", 00:31:16.294 "superblock": true, 00:31:16.294 "num_base_bdevs": 2, 00:31:16.294 "num_base_bdevs_discovered": 2, 00:31:16.294 "num_base_bdevs_operational": 2, 00:31:16.294 "base_bdevs_list": [ 00:31:16.294 { 00:31:16.294 "name": "spare", 00:31:16.294 "uuid": "ce0a4bcc-0059-54cb-bd4a-3455b477408c", 00:31:16.294 "is_configured": true, 00:31:16.294 "data_offset": 2048, 00:31:16.294 "data_size": 63488 00:31:16.294 }, 00:31:16.294 { 00:31:16.294 "name": "BaseBdev2", 00:31:16.294 "uuid": "80ea6115-d09e-573d-8300-abe8f8bd1084", 00:31:16.294 "is_configured": true, 00:31:16.294 "data_offset": 2048, 00:31:16.294 "data_size": 63488 00:31:16.294 } 00:31:16.294 ] 00:31:16.294 }' 00:31:16.294 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:16.294 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:16.552 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:31:16.552 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:16.552 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:31:16.552 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:31:16.552 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:16.552 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:16.552 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:16.552 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:16.552 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:16.552 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:16.552 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:16.552 "name": "raid_bdev1", 00:31:16.552 "uuid": "5100785b-7326-4f0a-9446-6e93c7dce5da", 00:31:16.552 "strip_size_kb": 0, 00:31:16.552 "state": "online", 00:31:16.552 "raid_level": "raid1", 00:31:16.552 "superblock": true, 00:31:16.552 "num_base_bdevs": 2, 00:31:16.552 "num_base_bdevs_discovered": 2, 00:31:16.552 "num_base_bdevs_operational": 2, 00:31:16.552 "base_bdevs_list": [ 00:31:16.552 { 00:31:16.552 "name": "spare", 00:31:16.552 "uuid": "ce0a4bcc-0059-54cb-bd4a-3455b477408c", 00:31:16.552 "is_configured": true, 00:31:16.552 "data_offset": 2048, 00:31:16.552 "data_size": 63488 00:31:16.552 }, 00:31:16.552 { 00:31:16.552 "name": "BaseBdev2", 00:31:16.552 "uuid": "80ea6115-d09e-573d-8300-abe8f8bd1084", 00:31:16.552 "is_configured": true, 00:31:16.552 "data_offset": 2048, 00:31:16.552 "data_size": 63488 00:31:16.552 } 00:31:16.552 ] 00:31:16.552 }' 00:31:16.552 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:16.552 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:31:16.552 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:16.552 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:31:16.552 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:16.552 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:31:16.552 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:16.552 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:16.552 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:16.810 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:31:16.810 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:31:16.810 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:16.810 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:16.810 [2024-11-20 13:51:23.634006] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:31:16.810 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:16.810 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:31:16.810 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:16.810 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:16.810 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:16.810 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:16.810 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:31:16.810 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:16.810 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:16.810 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:16.810 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:16.810 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:16.810 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:16.810 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:16.810 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:16.810 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:16.810 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:16.810 "name": "raid_bdev1", 00:31:16.810 "uuid": "5100785b-7326-4f0a-9446-6e93c7dce5da", 00:31:16.810 "strip_size_kb": 0, 00:31:16.810 "state": "online", 00:31:16.810 "raid_level": "raid1", 00:31:16.810 "superblock": true, 00:31:16.810 "num_base_bdevs": 2, 00:31:16.810 "num_base_bdevs_discovered": 1, 00:31:16.810 "num_base_bdevs_operational": 1, 00:31:16.810 "base_bdevs_list": [ 00:31:16.810 { 00:31:16.810 "name": null, 00:31:16.810 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:16.810 "is_configured": false, 00:31:16.810 "data_offset": 0, 00:31:16.810 "data_size": 63488 00:31:16.810 }, 00:31:16.810 { 00:31:16.810 "name": "BaseBdev2", 00:31:16.810 "uuid": "80ea6115-d09e-573d-8300-abe8f8bd1084", 00:31:16.810 "is_configured": true, 00:31:16.810 "data_offset": 2048, 00:31:16.810 "data_size": 63488 00:31:16.810 } 00:31:16.810 ] 00:31:16.810 }' 00:31:16.810 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:16.810 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:17.068 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:31:17.068 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:17.068 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:17.068 [2024-11-20 13:51:23.954096] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:31:17.068 [2024-11-20 13:51:23.954402] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:31:17.068 [2024-11-20 13:51:23.954489] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:31:17.068 [2024-11-20 13:51:23.954567] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:31:17.068 [2024-11-20 13:51:23.964602] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b1a0 00:31:17.068 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:17.068 13:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@757 -- # sleep 1 00:31:17.068 [2024-11-20 13:51:23.966345] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:31:18.017 13:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:18.017 13:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:18.017 13:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:18.017 13:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:18.017 13:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:18.017 13:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:18.017 13:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:18.017 13:51:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:18.017 13:51:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:18.017 13:51:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:18.017 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:18.017 "name": "raid_bdev1", 00:31:18.017 "uuid": "5100785b-7326-4f0a-9446-6e93c7dce5da", 00:31:18.017 "strip_size_kb": 0, 00:31:18.017 "state": "online", 00:31:18.017 "raid_level": "raid1", 00:31:18.017 "superblock": true, 00:31:18.017 "num_base_bdevs": 2, 00:31:18.017 "num_base_bdevs_discovered": 2, 00:31:18.017 "num_base_bdevs_operational": 2, 00:31:18.017 "process": { 00:31:18.017 "type": "rebuild", 00:31:18.017 "target": "spare", 00:31:18.017 "progress": { 00:31:18.017 "blocks": 20480, 00:31:18.017 "percent": 32 00:31:18.017 } 00:31:18.017 }, 00:31:18.017 "base_bdevs_list": [ 00:31:18.017 { 00:31:18.017 "name": "spare", 00:31:18.017 "uuid": "ce0a4bcc-0059-54cb-bd4a-3455b477408c", 00:31:18.017 "is_configured": true, 00:31:18.017 "data_offset": 2048, 00:31:18.017 "data_size": 63488 00:31:18.017 }, 00:31:18.017 { 00:31:18.017 "name": "BaseBdev2", 00:31:18.017 "uuid": "80ea6115-d09e-573d-8300-abe8f8bd1084", 00:31:18.017 "is_configured": true, 00:31:18.017 "data_offset": 2048, 00:31:18.017 "data_size": 63488 00:31:18.017 } 00:31:18.017 ] 00:31:18.017 }' 00:31:18.017 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:18.017 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:18.017 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:18.275 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:18.275 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:31:18.275 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:18.275 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:18.275 [2024-11-20 13:51:25.080615] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:31:18.275 [2024-11-20 13:51:25.173678] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:31:18.275 [2024-11-20 13:51:25.173764] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:18.275 [2024-11-20 13:51:25.173780] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:31:18.275 [2024-11-20 13:51:25.173787] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:31:18.275 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:18.275 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:31:18.275 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:18.275 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:18.275 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:18.275 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:18.275 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:31:18.275 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:18.275 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:18.275 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:18.275 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:18.275 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:18.275 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:18.275 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:18.275 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:18.275 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:18.275 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:18.275 "name": "raid_bdev1", 00:31:18.275 "uuid": "5100785b-7326-4f0a-9446-6e93c7dce5da", 00:31:18.275 "strip_size_kb": 0, 00:31:18.275 "state": "online", 00:31:18.275 "raid_level": "raid1", 00:31:18.275 "superblock": true, 00:31:18.275 "num_base_bdevs": 2, 00:31:18.275 "num_base_bdevs_discovered": 1, 00:31:18.275 "num_base_bdevs_operational": 1, 00:31:18.275 "base_bdevs_list": [ 00:31:18.275 { 00:31:18.275 "name": null, 00:31:18.275 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:18.275 "is_configured": false, 00:31:18.275 "data_offset": 0, 00:31:18.275 "data_size": 63488 00:31:18.275 }, 00:31:18.275 { 00:31:18.275 "name": "BaseBdev2", 00:31:18.275 "uuid": "80ea6115-d09e-573d-8300-abe8f8bd1084", 00:31:18.275 "is_configured": true, 00:31:18.275 "data_offset": 2048, 00:31:18.275 "data_size": 63488 00:31:18.275 } 00:31:18.275 ] 00:31:18.275 }' 00:31:18.275 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:18.275 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:18.533 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:31:18.533 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:18.533 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:18.533 [2024-11-20 13:51:25.495929] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:31:18.533 [2024-11-20 13:51:25.496009] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:18.533 [2024-11-20 13:51:25.496032] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:31:18.533 [2024-11-20 13:51:25.496041] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:18.533 [2024-11-20 13:51:25.496506] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:18.533 [2024-11-20 13:51:25.496525] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:31:18.533 [2024-11-20 13:51:25.496614] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:31:18.533 [2024-11-20 13:51:25.496626] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:31:18.533 [2024-11-20 13:51:25.496638] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:31:18.533 [2024-11-20 13:51:25.496655] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:31:18.533 [2024-11-20 13:51:25.506659] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b270 00:31:18.533 spare 00:31:18.533 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:18.533 13:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@764 -- # sleep 1 00:31:18.533 [2024-11-20 13:51:25.508495] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:31:19.464 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:19.464 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:19.464 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:19.464 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:19.464 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:19.464 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:19.464 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:19.464 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:19.464 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:19.723 "name": "raid_bdev1", 00:31:19.723 "uuid": "5100785b-7326-4f0a-9446-6e93c7dce5da", 00:31:19.723 "strip_size_kb": 0, 00:31:19.723 "state": "online", 00:31:19.723 "raid_level": "raid1", 00:31:19.723 "superblock": true, 00:31:19.723 "num_base_bdevs": 2, 00:31:19.723 "num_base_bdevs_discovered": 2, 00:31:19.723 "num_base_bdevs_operational": 2, 00:31:19.723 "process": { 00:31:19.723 "type": "rebuild", 00:31:19.723 "target": "spare", 00:31:19.723 "progress": { 00:31:19.723 "blocks": 20480, 00:31:19.723 "percent": 32 00:31:19.723 } 00:31:19.723 }, 00:31:19.723 "base_bdevs_list": [ 00:31:19.723 { 00:31:19.723 "name": "spare", 00:31:19.723 "uuid": "ce0a4bcc-0059-54cb-bd4a-3455b477408c", 00:31:19.723 "is_configured": true, 00:31:19.723 "data_offset": 2048, 00:31:19.723 "data_size": 63488 00:31:19.723 }, 00:31:19.723 { 00:31:19.723 "name": "BaseBdev2", 00:31:19.723 "uuid": "80ea6115-d09e-573d-8300-abe8f8bd1084", 00:31:19.723 "is_configured": true, 00:31:19.723 "data_offset": 2048, 00:31:19.723 "data_size": 63488 00:31:19.723 } 00:31:19.723 ] 00:31:19.723 }' 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:19.723 [2024-11-20 13:51:26.622735] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:31:19.723 [2024-11-20 13:51:26.715777] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:31:19.723 [2024-11-20 13:51:26.715868] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:19.723 [2024-11-20 13:51:26.715882] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:31:19.723 [2024-11-20 13:51:26.715890] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:19.723 "name": "raid_bdev1", 00:31:19.723 "uuid": "5100785b-7326-4f0a-9446-6e93c7dce5da", 00:31:19.723 "strip_size_kb": 0, 00:31:19.723 "state": "online", 00:31:19.723 "raid_level": "raid1", 00:31:19.723 "superblock": true, 00:31:19.723 "num_base_bdevs": 2, 00:31:19.723 "num_base_bdevs_discovered": 1, 00:31:19.723 "num_base_bdevs_operational": 1, 00:31:19.723 "base_bdevs_list": [ 00:31:19.723 { 00:31:19.723 "name": null, 00:31:19.723 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:19.723 "is_configured": false, 00:31:19.723 "data_offset": 0, 00:31:19.723 "data_size": 63488 00:31:19.723 }, 00:31:19.723 { 00:31:19.723 "name": "BaseBdev2", 00:31:19.723 "uuid": "80ea6115-d09e-573d-8300-abe8f8bd1084", 00:31:19.723 "is_configured": true, 00:31:19.723 "data_offset": 2048, 00:31:19.723 "data_size": 63488 00:31:19.723 } 00:31:19.723 ] 00:31:19.723 }' 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:19.723 13:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:20.289 13:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:31:20.289 13:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:20.289 13:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:31:20.289 13:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:31:20.289 13:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:20.289 13:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:20.290 13:51:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:20.290 13:51:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:20.290 13:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:20.290 13:51:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:20.290 13:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:20.290 "name": "raid_bdev1", 00:31:20.290 "uuid": "5100785b-7326-4f0a-9446-6e93c7dce5da", 00:31:20.290 "strip_size_kb": 0, 00:31:20.290 "state": "online", 00:31:20.290 "raid_level": "raid1", 00:31:20.290 "superblock": true, 00:31:20.290 "num_base_bdevs": 2, 00:31:20.290 "num_base_bdevs_discovered": 1, 00:31:20.290 "num_base_bdevs_operational": 1, 00:31:20.290 "base_bdevs_list": [ 00:31:20.290 { 00:31:20.290 "name": null, 00:31:20.290 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:20.290 "is_configured": false, 00:31:20.290 "data_offset": 0, 00:31:20.290 "data_size": 63488 00:31:20.290 }, 00:31:20.290 { 00:31:20.290 "name": "BaseBdev2", 00:31:20.290 "uuid": "80ea6115-d09e-573d-8300-abe8f8bd1084", 00:31:20.290 "is_configured": true, 00:31:20.290 "data_offset": 2048, 00:31:20.290 "data_size": 63488 00:31:20.290 } 00:31:20.290 ] 00:31:20.290 }' 00:31:20.290 13:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:20.290 13:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:31:20.290 13:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:20.290 13:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:31:20.290 13:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:31:20.290 13:51:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:20.290 13:51:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:20.290 13:51:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:20.290 13:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:31:20.290 13:51:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:20.290 13:51:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:20.290 [2024-11-20 13:51:27.161364] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:31:20.290 [2024-11-20 13:51:27.161529] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:20.290 [2024-11-20 13:51:27.161562] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:31:20.290 [2024-11-20 13:51:27.161577] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:20.290 [2024-11-20 13:51:27.162017] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:20.290 [2024-11-20 13:51:27.162032] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:31:20.290 [2024-11-20 13:51:27.162106] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:31:20.290 [2024-11-20 13:51:27.162124] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:31:20.290 [2024-11-20 13:51:27.162134] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:31:20.290 [2024-11-20 13:51:27.162144] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:31:20.290 BaseBdev1 00:31:20.290 13:51:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:20.290 13:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@775 -- # sleep 1 00:31:21.233 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:31:21.233 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:21.233 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:21.233 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:21.233 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:21.233 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:31:21.233 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:21.233 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:21.233 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:21.233 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:21.233 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:21.233 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:21.233 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:21.233 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:21.233 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:21.233 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:21.233 "name": "raid_bdev1", 00:31:21.233 "uuid": "5100785b-7326-4f0a-9446-6e93c7dce5da", 00:31:21.233 "strip_size_kb": 0, 00:31:21.233 "state": "online", 00:31:21.233 "raid_level": "raid1", 00:31:21.233 "superblock": true, 00:31:21.233 "num_base_bdevs": 2, 00:31:21.233 "num_base_bdevs_discovered": 1, 00:31:21.233 "num_base_bdevs_operational": 1, 00:31:21.233 "base_bdevs_list": [ 00:31:21.233 { 00:31:21.233 "name": null, 00:31:21.233 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:21.233 "is_configured": false, 00:31:21.233 "data_offset": 0, 00:31:21.233 "data_size": 63488 00:31:21.233 }, 00:31:21.233 { 00:31:21.233 "name": "BaseBdev2", 00:31:21.233 "uuid": "80ea6115-d09e-573d-8300-abe8f8bd1084", 00:31:21.233 "is_configured": true, 00:31:21.233 "data_offset": 2048, 00:31:21.233 "data_size": 63488 00:31:21.233 } 00:31:21.233 ] 00:31:21.233 }' 00:31:21.233 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:21.233 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:21.491 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:31:21.491 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:21.491 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:31:21.491 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:31:21.491 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:21.491 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:21.491 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:21.491 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:21.491 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:21.491 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:21.491 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:21.491 "name": "raid_bdev1", 00:31:21.491 "uuid": "5100785b-7326-4f0a-9446-6e93c7dce5da", 00:31:21.491 "strip_size_kb": 0, 00:31:21.491 "state": "online", 00:31:21.491 "raid_level": "raid1", 00:31:21.491 "superblock": true, 00:31:21.491 "num_base_bdevs": 2, 00:31:21.491 "num_base_bdevs_discovered": 1, 00:31:21.491 "num_base_bdevs_operational": 1, 00:31:21.491 "base_bdevs_list": [ 00:31:21.491 { 00:31:21.491 "name": null, 00:31:21.491 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:21.491 "is_configured": false, 00:31:21.491 "data_offset": 0, 00:31:21.491 "data_size": 63488 00:31:21.491 }, 00:31:21.491 { 00:31:21.491 "name": "BaseBdev2", 00:31:21.491 "uuid": "80ea6115-d09e-573d-8300-abe8f8bd1084", 00:31:21.491 "is_configured": true, 00:31:21.491 "data_offset": 2048, 00:31:21.491 "data_size": 63488 00:31:21.491 } 00:31:21.491 ] 00:31:21.491 }' 00:31:21.491 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:21.749 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:31:21.749 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:21.749 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:31:21.749 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:31:21.749 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@652 -- # local es=0 00:31:21.749 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:31:21.749 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:31:21.749 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:31:21.749 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:31:21.749 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:31:21.749 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:31:21.749 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:21.749 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:21.749 [2024-11-20 13:51:28.613848] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:31:21.749 [2024-11-20 13:51:28.614034] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:31:21.749 [2024-11-20 13:51:28.614046] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:31:21.749 request: 00:31:21.749 { 00:31:21.749 "base_bdev": "BaseBdev1", 00:31:21.749 "raid_bdev": "raid_bdev1", 00:31:21.749 "method": "bdev_raid_add_base_bdev", 00:31:21.749 "req_id": 1 00:31:21.749 } 00:31:21.749 Got JSON-RPC error response 00:31:21.749 response: 00:31:21.749 { 00:31:21.749 "code": -22, 00:31:21.749 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:31:21.749 } 00:31:21.749 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:31:21.749 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # es=1 00:31:21.749 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:31:21.749 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:31:21.749 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:31:21.749 13:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@779 -- # sleep 1 00:31:22.682 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:31:22.682 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:22.682 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:22.682 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:22.682 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:22.682 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:31:22.682 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:22.682 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:22.682 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:22.682 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:22.682 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:22.682 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:22.682 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:22.682 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:22.682 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:22.682 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:22.682 "name": "raid_bdev1", 00:31:22.682 "uuid": "5100785b-7326-4f0a-9446-6e93c7dce5da", 00:31:22.682 "strip_size_kb": 0, 00:31:22.682 "state": "online", 00:31:22.682 "raid_level": "raid1", 00:31:22.682 "superblock": true, 00:31:22.682 "num_base_bdevs": 2, 00:31:22.682 "num_base_bdevs_discovered": 1, 00:31:22.682 "num_base_bdevs_operational": 1, 00:31:22.682 "base_bdevs_list": [ 00:31:22.682 { 00:31:22.682 "name": null, 00:31:22.682 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:22.682 "is_configured": false, 00:31:22.682 "data_offset": 0, 00:31:22.682 "data_size": 63488 00:31:22.682 }, 00:31:22.682 { 00:31:22.682 "name": "BaseBdev2", 00:31:22.682 "uuid": "80ea6115-d09e-573d-8300-abe8f8bd1084", 00:31:22.682 "is_configured": true, 00:31:22.682 "data_offset": 2048, 00:31:22.682 "data_size": 63488 00:31:22.682 } 00:31:22.682 ] 00:31:22.682 }' 00:31:22.682 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:22.682 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:22.940 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:31:22.940 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:22.940 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:31:22.940 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:31:22.940 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:22.940 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:22.940 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:22.940 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:22.940 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:22.940 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:22.940 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:22.940 "name": "raid_bdev1", 00:31:22.940 "uuid": "5100785b-7326-4f0a-9446-6e93c7dce5da", 00:31:22.940 "strip_size_kb": 0, 00:31:22.940 "state": "online", 00:31:22.940 "raid_level": "raid1", 00:31:22.940 "superblock": true, 00:31:22.940 "num_base_bdevs": 2, 00:31:22.940 "num_base_bdevs_discovered": 1, 00:31:22.940 "num_base_bdevs_operational": 1, 00:31:22.940 "base_bdevs_list": [ 00:31:22.940 { 00:31:22.940 "name": null, 00:31:22.940 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:22.940 "is_configured": false, 00:31:22.940 "data_offset": 0, 00:31:22.940 "data_size": 63488 00:31:22.940 }, 00:31:22.940 { 00:31:22.940 "name": "BaseBdev2", 00:31:22.940 "uuid": "80ea6115-d09e-573d-8300-abe8f8bd1084", 00:31:22.940 "is_configured": true, 00:31:22.940 "data_offset": 2048, 00:31:22.940 "data_size": 63488 00:31:22.940 } 00:31:22.940 ] 00:31:22.940 }' 00:31:22.940 13:51:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:23.197 13:51:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:31:23.197 13:51:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:23.197 13:51:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:31:23.197 13:51:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@784 -- # killprocess 74699 00:31:23.197 13:51:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@954 -- # '[' -z 74699 ']' 00:31:23.197 13:51:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@958 -- # kill -0 74699 00:31:23.197 13:51:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # uname 00:31:23.197 13:51:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:31:23.197 13:51:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 74699 00:31:23.197 killing process with pid 74699 00:31:23.197 Received shutdown signal, test time was about 15.012984 seconds 00:31:23.197 00:31:23.197 Latency(us) 00:31:23.197 [2024-11-20T13:51:30.256Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:31:23.197 [2024-11-20T13:51:30.256Z] =================================================================================================================== 00:31:23.197 [2024-11-20T13:51:30.256Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:31:23.197 13:51:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:31:23.198 13:51:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:31:23.198 13:51:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 74699' 00:31:23.198 13:51:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@973 -- # kill 74699 00:31:23.198 [2024-11-20 13:51:30.064100] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:31:23.198 13:51:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@978 -- # wait 74699 00:31:23.198 [2024-11-20 13:51:30.064231] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:31:23.198 [2024-11-20 13:51:30.064282] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:31:23.198 [2024-11-20 13:51:30.064291] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:31:23.198 [2024-11-20 13:51:30.184515] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:31:23.826 13:51:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@786 -- # return 0 00:31:23.826 00:31:23.826 real 0m17.339s 00:31:23.826 user 0m21.991s 00:31:23.826 sys 0m1.587s 00:31:23.826 13:51:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:31:23.826 ************************************ 00:31:23.826 END TEST raid_rebuild_test_sb_io 00:31:23.826 ************************************ 00:31:23.826 13:51:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:31:24.100 13:51:30 bdev_raid -- bdev/bdev_raid.sh@977 -- # for n in 2 4 00:31:24.100 13:51:30 bdev_raid -- bdev/bdev_raid.sh@978 -- # run_test raid_rebuild_test raid_rebuild_test raid1 4 false false true 00:31:24.100 13:51:30 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:31:24.100 13:51:30 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:31:24.100 13:51:30 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:31:24.100 ************************************ 00:31:24.100 START TEST raid_rebuild_test 00:31:24.100 ************************************ 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 false false true 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:31:24.100 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=75361 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 75361 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 75361 ']' 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:24.100 13:51:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:31:24.100 I/O size of 3145728 is greater than zero copy threshold (65536). 00:31:24.100 Zero copy mechanism will not be used. 00:31:24.100 [2024-11-20 13:51:30.946336] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:31:24.100 [2024-11-20 13:51:30.946457] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid75361 ] 00:31:24.100 [2024-11-20 13:51:31.104725] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:31:24.359 [2024-11-20 13:51:31.225846] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:31:24.359 [2024-11-20 13:51:31.375115] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:31:24.359 [2024-11-20 13:51:31.375352] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:24.925 BaseBdev1_malloc 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:24.925 [2024-11-20 13:51:31.802642] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:31:24.925 [2024-11-20 13:51:31.802868] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:24.925 [2024-11-20 13:51:31.802898] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:31:24.925 [2024-11-20 13:51:31.802911] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:24.925 [2024-11-20 13:51:31.805218] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:24.925 [2024-11-20 13:51:31.805256] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:31:24.925 BaseBdev1 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:24.925 BaseBdev2_malloc 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:24.925 [2024-11-20 13:51:31.845155] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:31:24.925 [2024-11-20 13:51:31.845382] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:24.925 [2024-11-20 13:51:31.845410] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:31:24.925 [2024-11-20 13:51:31.845423] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:24.925 [2024-11-20 13:51:31.847645] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:24.925 [2024-11-20 13:51:31.847680] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:31:24.925 BaseBdev2 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:24.925 BaseBdev3_malloc 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:24.925 [2024-11-20 13:51:31.891761] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:31:24.925 [2024-11-20 13:51:31.891829] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:24.925 [2024-11-20 13:51:31.891852] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:31:24.925 [2024-11-20 13:51:31.891863] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:24.925 [2024-11-20 13:51:31.894097] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:24.925 [2024-11-20 13:51:31.894133] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:31:24.925 BaseBdev3 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:24.925 BaseBdev4_malloc 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:24.925 [2024-11-20 13:51:31.933959] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:31:24.925 [2024-11-20 13:51:31.934037] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:24.925 [2024-11-20 13:51:31.934058] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:31:24.925 [2024-11-20 13:51:31.934070] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:24.925 [2024-11-20 13:51:31.936274] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:24.925 [2024-11-20 13:51:31.936433] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:31:24.925 BaseBdev4 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:24.925 spare_malloc 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:24.925 spare_delay 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:24.925 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:24.925 [2024-11-20 13:51:31.979824] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:31:24.925 [2024-11-20 13:51:31.980032] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:24.925 [2024-11-20 13:51:31.980057] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:31:24.925 [2024-11-20 13:51:31.980070] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:25.184 [2024-11-20 13:51:31.982285] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:25.184 [2024-11-20 13:51:31.982319] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:31:25.184 spare 00:31:25.184 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:25.184 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:31:25.184 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:25.184 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:25.184 [2024-11-20 13:51:31.987873] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:31:25.184 [2024-11-20 13:51:31.989906] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:31:25.184 [2024-11-20 13:51:31.989968] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:31:25.184 [2024-11-20 13:51:31.990039] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:31:25.184 [2024-11-20 13:51:31.990121] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:31:25.184 [2024-11-20 13:51:31.990135] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:31:25.184 [2024-11-20 13:51:31.990401] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:31:25.184 [2024-11-20 13:51:31.990563] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:31:25.184 [2024-11-20 13:51:31.990574] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:31:25.184 [2024-11-20 13:51:31.990716] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:25.184 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:25.184 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:31:25.184 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:25.184 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:25.184 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:25.184 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:25.184 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:31:25.184 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:25.184 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:25.184 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:25.184 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:25.184 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:25.184 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:25.184 13:51:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:25.184 13:51:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:25.184 13:51:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:25.184 13:51:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:25.184 "name": "raid_bdev1", 00:31:25.184 "uuid": "d7a8ea31-d7f9-49dc-9c1b-b3bf5190caf8", 00:31:25.184 "strip_size_kb": 0, 00:31:25.184 "state": "online", 00:31:25.184 "raid_level": "raid1", 00:31:25.184 "superblock": false, 00:31:25.184 "num_base_bdevs": 4, 00:31:25.184 "num_base_bdevs_discovered": 4, 00:31:25.184 "num_base_bdevs_operational": 4, 00:31:25.184 "base_bdevs_list": [ 00:31:25.184 { 00:31:25.184 "name": "BaseBdev1", 00:31:25.184 "uuid": "02651123-2a99-5370-ac56-e0d9cbb01bad", 00:31:25.184 "is_configured": true, 00:31:25.184 "data_offset": 0, 00:31:25.184 "data_size": 65536 00:31:25.184 }, 00:31:25.184 { 00:31:25.184 "name": "BaseBdev2", 00:31:25.184 "uuid": "ff62aa2a-1900-572d-8d20-baab26694987", 00:31:25.184 "is_configured": true, 00:31:25.184 "data_offset": 0, 00:31:25.184 "data_size": 65536 00:31:25.184 }, 00:31:25.184 { 00:31:25.184 "name": "BaseBdev3", 00:31:25.184 "uuid": "433efd1d-99bb-5cee-8561-a05fcc4befed", 00:31:25.184 "is_configured": true, 00:31:25.184 "data_offset": 0, 00:31:25.184 "data_size": 65536 00:31:25.184 }, 00:31:25.184 { 00:31:25.184 "name": "BaseBdev4", 00:31:25.184 "uuid": "fa59b321-16d5-50ac-ace1-3b5324d204a5", 00:31:25.184 "is_configured": true, 00:31:25.184 "data_offset": 0, 00:31:25.184 "data_size": 65536 00:31:25.184 } 00:31:25.184 ] 00:31:25.184 }' 00:31:25.184 13:51:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:25.184 13:51:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:25.442 13:51:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:31:25.442 13:51:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:31:25.442 13:51:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:25.442 13:51:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:25.442 [2024-11-20 13:51:32.312345] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:31:25.442 13:51:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:25.442 13:51:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:31:25.442 13:51:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:25.442 13:51:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:31:25.442 13:51:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:25.442 13:51:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:25.442 13:51:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:25.442 13:51:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:31:25.442 13:51:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:31:25.442 13:51:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:31:25.442 13:51:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:31:25.442 13:51:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:31:25.442 13:51:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:31:25.442 13:51:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:31:25.442 13:51:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:31:25.442 13:51:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:31:25.442 13:51:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:31:25.442 13:51:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:31:25.442 13:51:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:31:25.442 13:51:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:31:25.442 13:51:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:31:25.700 [2024-11-20 13:51:32.556090] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:31:25.700 /dev/nbd0 00:31:25.700 13:51:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:31:25.700 13:51:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:31:25.700 13:51:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:31:25.700 13:51:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:31:25.700 13:51:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:31:25.700 13:51:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:31:25.700 13:51:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:31:25.700 13:51:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:31:25.700 13:51:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:31:25.700 13:51:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:31:25.700 13:51:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:31:25.700 1+0 records in 00:31:25.700 1+0 records out 00:31:25.700 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000382272 s, 10.7 MB/s 00:31:25.700 13:51:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:25.700 13:51:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:31:25.700 13:51:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:25.700 13:51:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:31:25.700 13:51:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:31:25.700 13:51:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:31:25.700 13:51:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:31:25.700 13:51:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:31:25.700 13:51:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:31:25.700 13:51:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=65536 oflag=direct 00:31:32.253 65536+0 records in 00:31:32.253 65536+0 records out 00:31:32.253 33554432 bytes (34 MB, 32 MiB) copied, 5.92374 s, 5.7 MB/s 00:31:32.253 13:51:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:31:32.253 13:51:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:31:32.253 13:51:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:31:32.253 13:51:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:31:32.253 13:51:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:31:32.253 13:51:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:31:32.253 13:51:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:31:32.253 [2024-11-20 13:51:38.732030] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:32.253 13:51:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:31:32.253 13:51:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:31:32.253 13:51:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:31:32.253 13:51:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:31:32.253 13:51:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:31:32.253 13:51:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:31:32.253 13:51:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:31:32.253 13:51:38 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:31:32.253 13:51:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:31:32.253 13:51:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:32.253 13:51:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:32.253 [2024-11-20 13:51:38.756550] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:31:32.253 13:51:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:32.253 13:51:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:31:32.253 13:51:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:32.254 13:51:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:32.254 13:51:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:32.254 13:51:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:32.254 13:51:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:31:32.254 13:51:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:32.254 13:51:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:32.254 13:51:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:32.254 13:51:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:32.254 13:51:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:32.254 13:51:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:32.254 13:51:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:32.254 13:51:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:32.254 13:51:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:32.254 13:51:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:32.254 "name": "raid_bdev1", 00:31:32.254 "uuid": "d7a8ea31-d7f9-49dc-9c1b-b3bf5190caf8", 00:31:32.254 "strip_size_kb": 0, 00:31:32.254 "state": "online", 00:31:32.254 "raid_level": "raid1", 00:31:32.254 "superblock": false, 00:31:32.254 "num_base_bdevs": 4, 00:31:32.254 "num_base_bdevs_discovered": 3, 00:31:32.254 "num_base_bdevs_operational": 3, 00:31:32.254 "base_bdevs_list": [ 00:31:32.254 { 00:31:32.254 "name": null, 00:31:32.254 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:32.254 "is_configured": false, 00:31:32.254 "data_offset": 0, 00:31:32.254 "data_size": 65536 00:31:32.254 }, 00:31:32.254 { 00:31:32.254 "name": "BaseBdev2", 00:31:32.254 "uuid": "ff62aa2a-1900-572d-8d20-baab26694987", 00:31:32.254 "is_configured": true, 00:31:32.254 "data_offset": 0, 00:31:32.254 "data_size": 65536 00:31:32.254 }, 00:31:32.254 { 00:31:32.254 "name": "BaseBdev3", 00:31:32.254 "uuid": "433efd1d-99bb-5cee-8561-a05fcc4befed", 00:31:32.254 "is_configured": true, 00:31:32.254 "data_offset": 0, 00:31:32.254 "data_size": 65536 00:31:32.254 }, 00:31:32.254 { 00:31:32.254 "name": "BaseBdev4", 00:31:32.254 "uuid": "fa59b321-16d5-50ac-ace1-3b5324d204a5", 00:31:32.254 "is_configured": true, 00:31:32.254 "data_offset": 0, 00:31:32.254 "data_size": 65536 00:31:32.254 } 00:31:32.254 ] 00:31:32.254 }' 00:31:32.254 13:51:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:32.254 13:51:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:32.254 13:51:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:31:32.254 13:51:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:32.254 13:51:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:32.254 [2024-11-20 13:51:39.092635] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:31:32.254 [2024-11-20 13:51:39.101352] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d09d70 00:31:32.254 13:51:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:32.254 13:51:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:31:32.254 [2024-11-20 13:51:39.103060] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:33.188 "name": "raid_bdev1", 00:31:33.188 "uuid": "d7a8ea31-d7f9-49dc-9c1b-b3bf5190caf8", 00:31:33.188 "strip_size_kb": 0, 00:31:33.188 "state": "online", 00:31:33.188 "raid_level": "raid1", 00:31:33.188 "superblock": false, 00:31:33.188 "num_base_bdevs": 4, 00:31:33.188 "num_base_bdevs_discovered": 4, 00:31:33.188 "num_base_bdevs_operational": 4, 00:31:33.188 "process": { 00:31:33.188 "type": "rebuild", 00:31:33.188 "target": "spare", 00:31:33.188 "progress": { 00:31:33.188 "blocks": 20480, 00:31:33.188 "percent": 31 00:31:33.188 } 00:31:33.188 }, 00:31:33.188 "base_bdevs_list": [ 00:31:33.188 { 00:31:33.188 "name": "spare", 00:31:33.188 "uuid": "c7149191-6a9d-5be6-a37b-cd014671ada3", 00:31:33.188 "is_configured": true, 00:31:33.188 "data_offset": 0, 00:31:33.188 "data_size": 65536 00:31:33.188 }, 00:31:33.188 { 00:31:33.188 "name": "BaseBdev2", 00:31:33.188 "uuid": "ff62aa2a-1900-572d-8d20-baab26694987", 00:31:33.188 "is_configured": true, 00:31:33.188 "data_offset": 0, 00:31:33.188 "data_size": 65536 00:31:33.188 }, 00:31:33.188 { 00:31:33.188 "name": "BaseBdev3", 00:31:33.188 "uuid": "433efd1d-99bb-5cee-8561-a05fcc4befed", 00:31:33.188 "is_configured": true, 00:31:33.188 "data_offset": 0, 00:31:33.188 "data_size": 65536 00:31:33.188 }, 00:31:33.188 { 00:31:33.188 "name": "BaseBdev4", 00:31:33.188 "uuid": "fa59b321-16d5-50ac-ace1-3b5324d204a5", 00:31:33.188 "is_configured": true, 00:31:33.188 "data_offset": 0, 00:31:33.188 "data_size": 65536 00:31:33.188 } 00:31:33.188 ] 00:31:33.188 }' 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:33.188 [2024-11-20 13:51:40.201098] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:31:33.188 [2024-11-20 13:51:40.209567] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:31:33.188 [2024-11-20 13:51:40.209625] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:33.188 [2024-11-20 13:51:40.209639] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:31:33.188 [2024-11-20 13:51:40.209648] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:33.188 13:51:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:33.189 13:51:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:33.447 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:33.447 "name": "raid_bdev1", 00:31:33.447 "uuid": "d7a8ea31-d7f9-49dc-9c1b-b3bf5190caf8", 00:31:33.447 "strip_size_kb": 0, 00:31:33.447 "state": "online", 00:31:33.447 "raid_level": "raid1", 00:31:33.447 "superblock": false, 00:31:33.447 "num_base_bdevs": 4, 00:31:33.447 "num_base_bdevs_discovered": 3, 00:31:33.447 "num_base_bdevs_operational": 3, 00:31:33.447 "base_bdevs_list": [ 00:31:33.447 { 00:31:33.447 "name": null, 00:31:33.447 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:33.447 "is_configured": false, 00:31:33.447 "data_offset": 0, 00:31:33.447 "data_size": 65536 00:31:33.447 }, 00:31:33.447 { 00:31:33.447 "name": "BaseBdev2", 00:31:33.447 "uuid": "ff62aa2a-1900-572d-8d20-baab26694987", 00:31:33.447 "is_configured": true, 00:31:33.447 "data_offset": 0, 00:31:33.447 "data_size": 65536 00:31:33.447 }, 00:31:33.447 { 00:31:33.447 "name": "BaseBdev3", 00:31:33.447 "uuid": "433efd1d-99bb-5cee-8561-a05fcc4befed", 00:31:33.447 "is_configured": true, 00:31:33.447 "data_offset": 0, 00:31:33.447 "data_size": 65536 00:31:33.447 }, 00:31:33.447 { 00:31:33.447 "name": "BaseBdev4", 00:31:33.447 "uuid": "fa59b321-16d5-50ac-ace1-3b5324d204a5", 00:31:33.447 "is_configured": true, 00:31:33.447 "data_offset": 0, 00:31:33.447 "data_size": 65536 00:31:33.447 } 00:31:33.447 ] 00:31:33.447 }' 00:31:33.447 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:33.447 13:51:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:33.705 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:31:33.705 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:33.705 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:31:33.705 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:31:33.705 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:33.705 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:33.705 13:51:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:33.705 13:51:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:33.705 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:33.705 13:51:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:33.705 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:33.705 "name": "raid_bdev1", 00:31:33.705 "uuid": "d7a8ea31-d7f9-49dc-9c1b-b3bf5190caf8", 00:31:33.705 "strip_size_kb": 0, 00:31:33.705 "state": "online", 00:31:33.705 "raid_level": "raid1", 00:31:33.705 "superblock": false, 00:31:33.705 "num_base_bdevs": 4, 00:31:33.705 "num_base_bdevs_discovered": 3, 00:31:33.705 "num_base_bdevs_operational": 3, 00:31:33.705 "base_bdevs_list": [ 00:31:33.705 { 00:31:33.705 "name": null, 00:31:33.705 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:33.705 "is_configured": false, 00:31:33.705 "data_offset": 0, 00:31:33.705 "data_size": 65536 00:31:33.705 }, 00:31:33.705 { 00:31:33.705 "name": "BaseBdev2", 00:31:33.705 "uuid": "ff62aa2a-1900-572d-8d20-baab26694987", 00:31:33.705 "is_configured": true, 00:31:33.705 "data_offset": 0, 00:31:33.705 "data_size": 65536 00:31:33.705 }, 00:31:33.705 { 00:31:33.705 "name": "BaseBdev3", 00:31:33.705 "uuid": "433efd1d-99bb-5cee-8561-a05fcc4befed", 00:31:33.705 "is_configured": true, 00:31:33.705 "data_offset": 0, 00:31:33.705 "data_size": 65536 00:31:33.705 }, 00:31:33.705 { 00:31:33.705 "name": "BaseBdev4", 00:31:33.705 "uuid": "fa59b321-16d5-50ac-ace1-3b5324d204a5", 00:31:33.705 "is_configured": true, 00:31:33.705 "data_offset": 0, 00:31:33.705 "data_size": 65536 00:31:33.705 } 00:31:33.705 ] 00:31:33.705 }' 00:31:33.705 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:33.705 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:31:33.705 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:33.705 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:31:33.705 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:31:33.705 13:51:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:33.705 13:51:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:33.705 [2024-11-20 13:51:40.646312] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:31:33.705 [2024-11-20 13:51:40.654535] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d09e40 00:31:33.705 13:51:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:33.705 13:51:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:31:33.705 [2024-11-20 13:51:40.656409] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:31:34.711 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:34.711 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:34.711 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:34.711 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:34.711 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:34.711 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:34.711 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:34.711 13:51:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:34.711 13:51:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:34.711 13:51:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:34.711 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:34.711 "name": "raid_bdev1", 00:31:34.711 "uuid": "d7a8ea31-d7f9-49dc-9c1b-b3bf5190caf8", 00:31:34.711 "strip_size_kb": 0, 00:31:34.711 "state": "online", 00:31:34.711 "raid_level": "raid1", 00:31:34.711 "superblock": false, 00:31:34.711 "num_base_bdevs": 4, 00:31:34.711 "num_base_bdevs_discovered": 4, 00:31:34.711 "num_base_bdevs_operational": 4, 00:31:34.711 "process": { 00:31:34.711 "type": "rebuild", 00:31:34.711 "target": "spare", 00:31:34.711 "progress": { 00:31:34.711 "blocks": 20480, 00:31:34.711 "percent": 31 00:31:34.711 } 00:31:34.711 }, 00:31:34.712 "base_bdevs_list": [ 00:31:34.712 { 00:31:34.712 "name": "spare", 00:31:34.712 "uuid": "c7149191-6a9d-5be6-a37b-cd014671ada3", 00:31:34.712 "is_configured": true, 00:31:34.712 "data_offset": 0, 00:31:34.712 "data_size": 65536 00:31:34.712 }, 00:31:34.712 { 00:31:34.712 "name": "BaseBdev2", 00:31:34.712 "uuid": "ff62aa2a-1900-572d-8d20-baab26694987", 00:31:34.712 "is_configured": true, 00:31:34.712 "data_offset": 0, 00:31:34.712 "data_size": 65536 00:31:34.712 }, 00:31:34.712 { 00:31:34.712 "name": "BaseBdev3", 00:31:34.712 "uuid": "433efd1d-99bb-5cee-8561-a05fcc4befed", 00:31:34.712 "is_configured": true, 00:31:34.712 "data_offset": 0, 00:31:34.712 "data_size": 65536 00:31:34.712 }, 00:31:34.712 { 00:31:34.712 "name": "BaseBdev4", 00:31:34.712 "uuid": "fa59b321-16d5-50ac-ace1-3b5324d204a5", 00:31:34.712 "is_configured": true, 00:31:34.712 "data_offset": 0, 00:31:34.712 "data_size": 65536 00:31:34.712 } 00:31:34.712 ] 00:31:34.712 }' 00:31:34.712 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:34.712 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:34.712 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:34.712 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:34.712 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:31:34.712 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:31:34.712 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:31:34.712 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:31:34.712 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:31:34.712 13:51:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:34.712 13:51:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:34.970 [2024-11-20 13:51:41.770296] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:31:34.970 [2024-11-20 13:51:41.863999] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000d09e40 00:31:34.970 13:51:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:34.970 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:31:34.970 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:31:34.970 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:34.970 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:34.970 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:34.970 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:34.970 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:34.970 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:34.970 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:34.970 13:51:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:34.970 13:51:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:34.970 13:51:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:34.970 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:34.970 "name": "raid_bdev1", 00:31:34.970 "uuid": "d7a8ea31-d7f9-49dc-9c1b-b3bf5190caf8", 00:31:34.970 "strip_size_kb": 0, 00:31:34.970 "state": "online", 00:31:34.970 "raid_level": "raid1", 00:31:34.970 "superblock": false, 00:31:34.970 "num_base_bdevs": 4, 00:31:34.970 "num_base_bdevs_discovered": 3, 00:31:34.970 "num_base_bdevs_operational": 3, 00:31:34.970 "process": { 00:31:34.970 "type": "rebuild", 00:31:34.970 "target": "spare", 00:31:34.970 "progress": { 00:31:34.970 "blocks": 24576, 00:31:34.970 "percent": 37 00:31:34.970 } 00:31:34.970 }, 00:31:34.970 "base_bdevs_list": [ 00:31:34.970 { 00:31:34.970 "name": "spare", 00:31:34.970 "uuid": "c7149191-6a9d-5be6-a37b-cd014671ada3", 00:31:34.970 "is_configured": true, 00:31:34.970 "data_offset": 0, 00:31:34.970 "data_size": 65536 00:31:34.970 }, 00:31:34.970 { 00:31:34.970 "name": null, 00:31:34.970 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:34.970 "is_configured": false, 00:31:34.970 "data_offset": 0, 00:31:34.970 "data_size": 65536 00:31:34.970 }, 00:31:34.970 { 00:31:34.970 "name": "BaseBdev3", 00:31:34.970 "uuid": "433efd1d-99bb-5cee-8561-a05fcc4befed", 00:31:34.970 "is_configured": true, 00:31:34.970 "data_offset": 0, 00:31:34.970 "data_size": 65536 00:31:34.970 }, 00:31:34.970 { 00:31:34.970 "name": "BaseBdev4", 00:31:34.970 "uuid": "fa59b321-16d5-50ac-ace1-3b5324d204a5", 00:31:34.970 "is_configured": true, 00:31:34.970 "data_offset": 0, 00:31:34.970 "data_size": 65536 00:31:34.970 } 00:31:34.971 ] 00:31:34.971 }' 00:31:34.971 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:34.971 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:34.971 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:34.971 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:34.971 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=348 00:31:34.971 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:34.971 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:34.971 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:34.971 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:34.971 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:34.971 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:34.971 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:34.971 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:34.971 13:51:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:34.971 13:51:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:34.971 13:51:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:34.971 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:34.971 "name": "raid_bdev1", 00:31:34.971 "uuid": "d7a8ea31-d7f9-49dc-9c1b-b3bf5190caf8", 00:31:34.971 "strip_size_kb": 0, 00:31:34.971 "state": "online", 00:31:34.971 "raid_level": "raid1", 00:31:34.971 "superblock": false, 00:31:34.971 "num_base_bdevs": 4, 00:31:34.971 "num_base_bdevs_discovered": 3, 00:31:34.971 "num_base_bdevs_operational": 3, 00:31:34.971 "process": { 00:31:34.971 "type": "rebuild", 00:31:34.971 "target": "spare", 00:31:34.971 "progress": { 00:31:34.971 "blocks": 26624, 00:31:34.971 "percent": 40 00:31:34.971 } 00:31:34.971 }, 00:31:34.971 "base_bdevs_list": [ 00:31:34.971 { 00:31:34.971 "name": "spare", 00:31:34.971 "uuid": "c7149191-6a9d-5be6-a37b-cd014671ada3", 00:31:34.971 "is_configured": true, 00:31:34.971 "data_offset": 0, 00:31:34.971 "data_size": 65536 00:31:34.971 }, 00:31:34.971 { 00:31:34.971 "name": null, 00:31:34.971 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:34.971 "is_configured": false, 00:31:34.971 "data_offset": 0, 00:31:34.971 "data_size": 65536 00:31:34.971 }, 00:31:34.971 { 00:31:34.971 "name": "BaseBdev3", 00:31:34.971 "uuid": "433efd1d-99bb-5cee-8561-a05fcc4befed", 00:31:34.971 "is_configured": true, 00:31:34.971 "data_offset": 0, 00:31:34.971 "data_size": 65536 00:31:34.971 }, 00:31:34.971 { 00:31:34.971 "name": "BaseBdev4", 00:31:34.971 "uuid": "fa59b321-16d5-50ac-ace1-3b5324d204a5", 00:31:34.971 "is_configured": true, 00:31:34.971 "data_offset": 0, 00:31:34.971 "data_size": 65536 00:31:34.971 } 00:31:34.971 ] 00:31:34.971 }' 00:31:34.971 13:51:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:35.229 13:51:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:35.229 13:51:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:35.229 13:51:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:35.229 13:51:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:31:36.162 13:51:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:36.162 13:51:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:36.162 13:51:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:36.162 13:51:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:36.162 13:51:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:36.162 13:51:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:36.162 13:51:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:36.162 13:51:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:36.162 13:51:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:36.162 13:51:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:36.162 13:51:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:36.162 13:51:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:36.162 "name": "raid_bdev1", 00:31:36.162 "uuid": "d7a8ea31-d7f9-49dc-9c1b-b3bf5190caf8", 00:31:36.162 "strip_size_kb": 0, 00:31:36.162 "state": "online", 00:31:36.162 "raid_level": "raid1", 00:31:36.162 "superblock": false, 00:31:36.162 "num_base_bdevs": 4, 00:31:36.162 "num_base_bdevs_discovered": 3, 00:31:36.162 "num_base_bdevs_operational": 3, 00:31:36.162 "process": { 00:31:36.162 "type": "rebuild", 00:31:36.162 "target": "spare", 00:31:36.162 "progress": { 00:31:36.162 "blocks": 49152, 00:31:36.162 "percent": 75 00:31:36.162 } 00:31:36.162 }, 00:31:36.162 "base_bdevs_list": [ 00:31:36.162 { 00:31:36.162 "name": "spare", 00:31:36.162 "uuid": "c7149191-6a9d-5be6-a37b-cd014671ada3", 00:31:36.162 "is_configured": true, 00:31:36.162 "data_offset": 0, 00:31:36.162 "data_size": 65536 00:31:36.162 }, 00:31:36.162 { 00:31:36.162 "name": null, 00:31:36.162 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:36.162 "is_configured": false, 00:31:36.162 "data_offset": 0, 00:31:36.162 "data_size": 65536 00:31:36.162 }, 00:31:36.162 { 00:31:36.162 "name": "BaseBdev3", 00:31:36.162 "uuid": "433efd1d-99bb-5cee-8561-a05fcc4befed", 00:31:36.162 "is_configured": true, 00:31:36.162 "data_offset": 0, 00:31:36.162 "data_size": 65536 00:31:36.162 }, 00:31:36.162 { 00:31:36.162 "name": "BaseBdev4", 00:31:36.162 "uuid": "fa59b321-16d5-50ac-ace1-3b5324d204a5", 00:31:36.162 "is_configured": true, 00:31:36.162 "data_offset": 0, 00:31:36.162 "data_size": 65536 00:31:36.162 } 00:31:36.162 ] 00:31:36.162 }' 00:31:36.162 13:51:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:36.162 13:51:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:36.162 13:51:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:36.162 13:51:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:36.162 13:51:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:31:37.095 [2024-11-20 13:51:43.876004] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:31:37.095 [2024-11-20 13:51:43.876240] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:31:37.095 [2024-11-20 13:51:43.876291] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:37.354 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:37.354 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:37.354 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:37.354 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:37.354 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:37.354 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:37.354 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:37.354 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:37.354 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:37.354 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:37.355 "name": "raid_bdev1", 00:31:37.355 "uuid": "d7a8ea31-d7f9-49dc-9c1b-b3bf5190caf8", 00:31:37.355 "strip_size_kb": 0, 00:31:37.355 "state": "online", 00:31:37.355 "raid_level": "raid1", 00:31:37.355 "superblock": false, 00:31:37.355 "num_base_bdevs": 4, 00:31:37.355 "num_base_bdevs_discovered": 3, 00:31:37.355 "num_base_bdevs_operational": 3, 00:31:37.355 "base_bdevs_list": [ 00:31:37.355 { 00:31:37.355 "name": "spare", 00:31:37.355 "uuid": "c7149191-6a9d-5be6-a37b-cd014671ada3", 00:31:37.355 "is_configured": true, 00:31:37.355 "data_offset": 0, 00:31:37.355 "data_size": 65536 00:31:37.355 }, 00:31:37.355 { 00:31:37.355 "name": null, 00:31:37.355 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:37.355 "is_configured": false, 00:31:37.355 "data_offset": 0, 00:31:37.355 "data_size": 65536 00:31:37.355 }, 00:31:37.355 { 00:31:37.355 "name": "BaseBdev3", 00:31:37.355 "uuid": "433efd1d-99bb-5cee-8561-a05fcc4befed", 00:31:37.355 "is_configured": true, 00:31:37.355 "data_offset": 0, 00:31:37.355 "data_size": 65536 00:31:37.355 }, 00:31:37.355 { 00:31:37.355 "name": "BaseBdev4", 00:31:37.355 "uuid": "fa59b321-16d5-50ac-ace1-3b5324d204a5", 00:31:37.355 "is_configured": true, 00:31:37.355 "data_offset": 0, 00:31:37.355 "data_size": 65536 00:31:37.355 } 00:31:37.355 ] 00:31:37.355 }' 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:37.355 "name": "raid_bdev1", 00:31:37.355 "uuid": "d7a8ea31-d7f9-49dc-9c1b-b3bf5190caf8", 00:31:37.355 "strip_size_kb": 0, 00:31:37.355 "state": "online", 00:31:37.355 "raid_level": "raid1", 00:31:37.355 "superblock": false, 00:31:37.355 "num_base_bdevs": 4, 00:31:37.355 "num_base_bdevs_discovered": 3, 00:31:37.355 "num_base_bdevs_operational": 3, 00:31:37.355 "base_bdevs_list": [ 00:31:37.355 { 00:31:37.355 "name": "spare", 00:31:37.355 "uuid": "c7149191-6a9d-5be6-a37b-cd014671ada3", 00:31:37.355 "is_configured": true, 00:31:37.355 "data_offset": 0, 00:31:37.355 "data_size": 65536 00:31:37.355 }, 00:31:37.355 { 00:31:37.355 "name": null, 00:31:37.355 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:37.355 "is_configured": false, 00:31:37.355 "data_offset": 0, 00:31:37.355 "data_size": 65536 00:31:37.355 }, 00:31:37.355 { 00:31:37.355 "name": "BaseBdev3", 00:31:37.355 "uuid": "433efd1d-99bb-5cee-8561-a05fcc4befed", 00:31:37.355 "is_configured": true, 00:31:37.355 "data_offset": 0, 00:31:37.355 "data_size": 65536 00:31:37.355 }, 00:31:37.355 { 00:31:37.355 "name": "BaseBdev4", 00:31:37.355 "uuid": "fa59b321-16d5-50ac-ace1-3b5324d204a5", 00:31:37.355 "is_configured": true, 00:31:37.355 "data_offset": 0, 00:31:37.355 "data_size": 65536 00:31:37.355 } 00:31:37.355 ] 00:31:37.355 }' 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:37.355 "name": "raid_bdev1", 00:31:37.355 "uuid": "d7a8ea31-d7f9-49dc-9c1b-b3bf5190caf8", 00:31:37.355 "strip_size_kb": 0, 00:31:37.355 "state": "online", 00:31:37.355 "raid_level": "raid1", 00:31:37.355 "superblock": false, 00:31:37.355 "num_base_bdevs": 4, 00:31:37.355 "num_base_bdevs_discovered": 3, 00:31:37.355 "num_base_bdevs_operational": 3, 00:31:37.355 "base_bdevs_list": [ 00:31:37.355 { 00:31:37.355 "name": "spare", 00:31:37.355 "uuid": "c7149191-6a9d-5be6-a37b-cd014671ada3", 00:31:37.355 "is_configured": true, 00:31:37.355 "data_offset": 0, 00:31:37.355 "data_size": 65536 00:31:37.355 }, 00:31:37.355 { 00:31:37.355 "name": null, 00:31:37.355 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:37.355 "is_configured": false, 00:31:37.355 "data_offset": 0, 00:31:37.355 "data_size": 65536 00:31:37.355 }, 00:31:37.355 { 00:31:37.355 "name": "BaseBdev3", 00:31:37.355 "uuid": "433efd1d-99bb-5cee-8561-a05fcc4befed", 00:31:37.355 "is_configured": true, 00:31:37.355 "data_offset": 0, 00:31:37.355 "data_size": 65536 00:31:37.355 }, 00:31:37.355 { 00:31:37.355 "name": "BaseBdev4", 00:31:37.355 "uuid": "fa59b321-16d5-50ac-ace1-3b5324d204a5", 00:31:37.355 "is_configured": true, 00:31:37.355 "data_offset": 0, 00:31:37.355 "data_size": 65536 00:31:37.355 } 00:31:37.355 ] 00:31:37.355 }' 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:37.355 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:37.921 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:31:37.921 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:37.921 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:37.921 [2024-11-20 13:51:44.709156] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:31:37.921 [2024-11-20 13:51:44.709320] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:31:37.921 [2024-11-20 13:51:44.709410] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:31:37.921 [2024-11-20 13:51:44.709489] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:31:37.921 [2024-11-20 13:51:44.709497] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:31:37.921 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:37.921 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:31:37.921 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:37.921 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:37.921 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:37.921 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:37.921 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:31:37.921 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:31:37.921 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:31:37.921 13:51:44 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:31:37.921 13:51:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:31:37.921 13:51:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:31:37.921 13:51:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:31:37.921 13:51:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:31:37.921 13:51:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:31:37.921 13:51:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:31:37.921 13:51:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:31:37.921 13:51:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:31:37.921 13:51:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:31:37.921 /dev/nbd0 00:31:37.921 13:51:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:31:38.180 13:51:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:31:38.180 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:31:38.180 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:31:38.180 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:31:38.180 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:31:38.180 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:31:38.180 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:31:38.180 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:31:38.180 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:31:38.180 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:31:38.180 1+0 records in 00:31:38.180 1+0 records out 00:31:38.180 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000260201 s, 15.7 MB/s 00:31:38.180 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:38.180 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:31:38.180 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:38.180 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:31:38.180 13:51:44 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:31:38.180 13:51:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:31:38.180 13:51:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:31:38.180 13:51:44 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:31:38.180 /dev/nbd1 00:31:38.180 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:31:38.180 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:31:38.180 13:51:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:31:38.180 13:51:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:31:38.180 13:51:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:31:38.180 13:51:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:31:38.180 13:51:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:31:38.180 13:51:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:31:38.180 13:51:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:31:38.180 13:51:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:31:38.180 13:51:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:31:38.180 1+0 records in 00:31:38.180 1+0 records out 00:31:38.180 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000296147 s, 13.8 MB/s 00:31:38.180 13:51:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:38.180 13:51:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:31:38.180 13:51:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:38.180 13:51:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:31:38.180 13:51:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:31:38.180 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:31:38.180 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:31:38.180 13:51:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:31:38.438 13:51:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:31:38.438 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:31:38.438 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:31:38.438 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:31:38.438 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:31:38.438 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:31:38.438 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:31:38.695 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:31:38.696 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:31:38.696 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:31:38.696 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:31:38.696 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:31:38.696 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:31:38.696 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:31:38.696 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:31:38.696 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:31:38.696 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:31:38.953 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:31:38.953 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:31:38.953 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:31:38.953 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:31:38.953 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:31:38.953 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:31:38.953 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:31:38.953 13:51:45 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:31:38.953 13:51:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:31:38.953 13:51:45 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 75361 00:31:38.953 13:51:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 75361 ']' 00:31:38.953 13:51:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 75361 00:31:38.953 13:51:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:31:38.953 13:51:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:31:38.953 13:51:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 75361 00:31:38.953 killing process with pid 75361 00:31:38.953 Received shutdown signal, test time was about 60.000000 seconds 00:31:38.953 00:31:38.953 Latency(us) 00:31:38.953 [2024-11-20T13:51:46.013Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:31:38.954 [2024-11-20T13:51:46.013Z] =================================================================================================================== 00:31:38.954 [2024-11-20T13:51:46.013Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:31:38.954 13:51:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:31:38.954 13:51:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:31:38.954 13:51:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 75361' 00:31:38.954 13:51:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@973 -- # kill 75361 00:31:38.954 [2024-11-20 13:51:45.821421] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:31:38.954 13:51:45 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@978 -- # wait 75361 00:31:39.211 [2024-11-20 13:51:46.137494] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:31:40.192 13:51:46 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:31:40.192 00:31:40.192 real 0m16.014s 00:31:40.192 user 0m17.436s 00:31:40.192 sys 0m2.740s 00:31:40.192 13:51:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:31:40.192 13:51:46 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:40.192 ************************************ 00:31:40.192 END TEST raid_rebuild_test 00:31:40.192 ************************************ 00:31:40.192 13:51:46 bdev_raid -- bdev/bdev_raid.sh@979 -- # run_test raid_rebuild_test_sb raid_rebuild_test raid1 4 true false true 00:31:40.192 13:51:46 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:31:40.192 13:51:46 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:31:40.192 13:51:46 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:31:40.192 ************************************ 00:31:40.192 START TEST raid_rebuild_test_sb 00:31:40.192 ************************************ 00:31:40.192 13:51:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 true false true 00:31:40.192 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:31:40.192 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:31:40.192 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:31:40.192 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:31:40.192 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:31:40.192 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:31:40.192 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:31:40.192 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:31:40.192 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:31:40.192 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:31:40.192 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:31:40.192 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:31:40.192 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:31:40.192 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:31:40.193 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:31:40.193 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:31:40.193 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:31:40.193 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:31:40.193 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:31:40.193 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:31:40.193 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:31:40.193 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:31:40.193 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:31:40.193 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:31:40.193 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:31:40.193 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:31:40.193 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:31:40.193 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:31:40.193 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:31:40.193 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:31:40.193 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:31:40.193 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=75797 00:31:40.193 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 75797 00:31:40.193 13:51:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 75797 ']' 00:31:40.193 13:51:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:31:40.193 13:51:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:31:40.193 13:51:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:31:40.193 13:51:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:31:40.193 13:51:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:40.193 13:51:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:31:40.193 I/O size of 3145728 is greater than zero copy threshold (65536). 00:31:40.193 Zero copy mechanism will not be used. 00:31:40.193 [2024-11-20 13:51:47.013454] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:31:40.193 [2024-11-20 13:51:47.013573] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid75797 ] 00:31:40.193 [2024-11-20 13:51:47.172474] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:31:40.451 [2024-11-20 13:51:47.289721] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:31:40.451 [2024-11-20 13:51:47.437788] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:31:40.451 [2024-11-20 13:51:47.437847] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:41.017 BaseBdev1_malloc 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:41.017 [2024-11-20 13:51:47.863776] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:31:41.017 [2024-11-20 13:51:47.863846] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:41.017 [2024-11-20 13:51:47.863870] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:31:41.017 [2024-11-20 13:51:47.863882] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:41.017 [2024-11-20 13:51:47.866225] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:41.017 [2024-11-20 13:51:47.866265] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:31:41.017 BaseBdev1 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:41.017 BaseBdev2_malloc 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:41.017 [2024-11-20 13:51:47.902334] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:31:41.017 [2024-11-20 13:51:47.902402] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:41.017 [2024-11-20 13:51:47.902424] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:31:41.017 [2024-11-20 13:51:47.902435] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:41.017 [2024-11-20 13:51:47.904726] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:41.017 [2024-11-20 13:51:47.904759] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:31:41.017 BaseBdev2 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:41.017 BaseBdev3_malloc 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:41.017 [2024-11-20 13:51:47.956072] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:31:41.017 [2024-11-20 13:51:47.956306] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:41.017 [2024-11-20 13:51:47.956336] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:31:41.017 [2024-11-20 13:51:47.956348] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:41.017 [2024-11-20 13:51:47.958655] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:41.017 [2024-11-20 13:51:47.958694] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:31:41.017 BaseBdev3 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:41.017 BaseBdev4_malloc 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:41.017 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:41.017 [2024-11-20 13:51:47.994021] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:31:41.018 [2024-11-20 13:51:47.994072] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:41.018 [2024-11-20 13:51:47.994089] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:31:41.018 [2024-11-20 13:51:47.994101] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:41.018 [2024-11-20 13:51:47.996305] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:41.018 [2024-11-20 13:51:47.996340] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:31:41.018 BaseBdev4 00:31:41.018 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:41.018 13:51:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:31:41.018 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:41.018 13:51:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:41.018 spare_malloc 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:41.018 spare_delay 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:41.018 [2024-11-20 13:51:48.040296] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:31:41.018 [2024-11-20 13:51:48.040358] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:41.018 [2024-11-20 13:51:48.040379] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:31:41.018 [2024-11-20 13:51:48.040390] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:41.018 [2024-11-20 13:51:48.042636] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:41.018 [2024-11-20 13:51:48.042674] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:31:41.018 spare 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:41.018 [2024-11-20 13:51:48.048352] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:31:41.018 [2024-11-20 13:51:48.050285] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:31:41.018 [2024-11-20 13:51:48.050347] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:31:41.018 [2024-11-20 13:51:48.050400] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:31:41.018 [2024-11-20 13:51:48.050583] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:31:41.018 [2024-11-20 13:51:48.050596] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:31:41.018 [2024-11-20 13:51:48.050860] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:31:41.018 [2024-11-20 13:51:48.051043] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:31:41.018 [2024-11-20 13:51:48.051054] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:31:41.018 [2024-11-20 13:51:48.051200] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:41.018 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:41.276 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:41.276 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:41.276 "name": "raid_bdev1", 00:31:41.276 "uuid": "befab01c-3204-49aa-9025-c8f88189ffd6", 00:31:41.276 "strip_size_kb": 0, 00:31:41.276 "state": "online", 00:31:41.276 "raid_level": "raid1", 00:31:41.276 "superblock": true, 00:31:41.276 "num_base_bdevs": 4, 00:31:41.276 "num_base_bdevs_discovered": 4, 00:31:41.276 "num_base_bdevs_operational": 4, 00:31:41.276 "base_bdevs_list": [ 00:31:41.276 { 00:31:41.276 "name": "BaseBdev1", 00:31:41.276 "uuid": "90add5aa-77b2-5116-81b7-6ee335efff0d", 00:31:41.276 "is_configured": true, 00:31:41.276 "data_offset": 2048, 00:31:41.276 "data_size": 63488 00:31:41.276 }, 00:31:41.276 { 00:31:41.276 "name": "BaseBdev2", 00:31:41.276 "uuid": "949fd081-c278-59f1-9c08-d54bc625023f", 00:31:41.276 "is_configured": true, 00:31:41.276 "data_offset": 2048, 00:31:41.276 "data_size": 63488 00:31:41.276 }, 00:31:41.276 { 00:31:41.276 "name": "BaseBdev3", 00:31:41.276 "uuid": "59eda477-1d7f-5ac2-9bfb-d02bcbc4c870", 00:31:41.276 "is_configured": true, 00:31:41.276 "data_offset": 2048, 00:31:41.276 "data_size": 63488 00:31:41.276 }, 00:31:41.276 { 00:31:41.276 "name": "BaseBdev4", 00:31:41.276 "uuid": "d885740a-b18e-5b70-a7fd-311240bd474e", 00:31:41.276 "is_configured": true, 00:31:41.276 "data_offset": 2048, 00:31:41.276 "data_size": 63488 00:31:41.276 } 00:31:41.276 ] 00:31:41.276 }' 00:31:41.276 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:41.276 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:41.534 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:31:41.534 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:31:41.534 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:41.534 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:41.534 [2024-11-20 13:51:48.376834] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:31:41.534 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:41.534 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:31:41.534 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:41.534 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:41.534 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:41.534 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:31:41.534 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:41.534 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:31:41.534 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:31:41.534 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:31:41.534 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:31:41.534 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:31:41.534 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:31:41.534 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:31:41.534 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:31:41.534 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:31:41.534 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:31:41.534 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:31:41.534 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:31:41.534 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:31:41.534 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:31:41.792 [2024-11-20 13:51:48.636590] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:31:41.792 /dev/nbd0 00:31:41.792 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:31:41.792 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:31:41.792 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:31:41.792 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:31:41.792 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:31:41.792 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:31:41.792 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:31:41.792 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:31:41.792 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:31:41.792 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:31:41.792 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:31:41.792 1+0 records in 00:31:41.792 1+0 records out 00:31:41.792 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00051699 s, 7.9 MB/s 00:31:41.792 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:41.792 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:31:41.792 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:41.792 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:31:41.792 13:51:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:31:41.792 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:31:41.792 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:31:41.792 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:31:41.792 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:31:41.792 13:51:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=63488 oflag=direct 00:31:48.345 63488+0 records in 00:31:48.345 63488+0 records out 00:31:48.345 32505856 bytes (33 MB, 31 MiB) copied, 5.8686 s, 5.5 MB/s 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:31:48.346 [2024-11-20 13:51:54.780191] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:48.346 [2024-11-20 13:51:54.794930] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:48.346 "name": "raid_bdev1", 00:31:48.346 "uuid": "befab01c-3204-49aa-9025-c8f88189ffd6", 00:31:48.346 "strip_size_kb": 0, 00:31:48.346 "state": "online", 00:31:48.346 "raid_level": "raid1", 00:31:48.346 "superblock": true, 00:31:48.346 "num_base_bdevs": 4, 00:31:48.346 "num_base_bdevs_discovered": 3, 00:31:48.346 "num_base_bdevs_operational": 3, 00:31:48.346 "base_bdevs_list": [ 00:31:48.346 { 00:31:48.346 "name": null, 00:31:48.346 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:48.346 "is_configured": false, 00:31:48.346 "data_offset": 0, 00:31:48.346 "data_size": 63488 00:31:48.346 }, 00:31:48.346 { 00:31:48.346 "name": "BaseBdev2", 00:31:48.346 "uuid": "949fd081-c278-59f1-9c08-d54bc625023f", 00:31:48.346 "is_configured": true, 00:31:48.346 "data_offset": 2048, 00:31:48.346 "data_size": 63488 00:31:48.346 }, 00:31:48.346 { 00:31:48.346 "name": "BaseBdev3", 00:31:48.346 "uuid": "59eda477-1d7f-5ac2-9bfb-d02bcbc4c870", 00:31:48.346 "is_configured": true, 00:31:48.346 "data_offset": 2048, 00:31:48.346 "data_size": 63488 00:31:48.346 }, 00:31:48.346 { 00:31:48.346 "name": "BaseBdev4", 00:31:48.346 "uuid": "d885740a-b18e-5b70-a7fd-311240bd474e", 00:31:48.346 "is_configured": true, 00:31:48.346 "data_offset": 2048, 00:31:48.346 "data_size": 63488 00:31:48.346 } 00:31:48.346 ] 00:31:48.346 }' 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:48.346 13:51:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:48.346 13:51:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:31:48.346 13:51:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:48.346 13:51:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:48.346 [2024-11-20 13:51:55.143018] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:31:48.346 [2024-11-20 13:51:55.151787] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000ca3500 00:31:48.346 13:51:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:48.346 13:51:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:31:48.346 [2024-11-20 13:51:55.153561] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:31:49.281 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:49.281 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:49.281 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:49.281 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:49.281 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:49.281 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:49.281 13:51:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:49.281 13:51:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:49.281 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:49.281 13:51:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:49.281 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:49.281 "name": "raid_bdev1", 00:31:49.281 "uuid": "befab01c-3204-49aa-9025-c8f88189ffd6", 00:31:49.281 "strip_size_kb": 0, 00:31:49.281 "state": "online", 00:31:49.281 "raid_level": "raid1", 00:31:49.281 "superblock": true, 00:31:49.281 "num_base_bdevs": 4, 00:31:49.281 "num_base_bdevs_discovered": 4, 00:31:49.281 "num_base_bdevs_operational": 4, 00:31:49.281 "process": { 00:31:49.281 "type": "rebuild", 00:31:49.281 "target": "spare", 00:31:49.281 "progress": { 00:31:49.281 "blocks": 20480, 00:31:49.281 "percent": 32 00:31:49.281 } 00:31:49.281 }, 00:31:49.281 "base_bdevs_list": [ 00:31:49.281 { 00:31:49.281 "name": "spare", 00:31:49.281 "uuid": "4c9ebeab-bb1e-5072-9065-732e55c929b5", 00:31:49.281 "is_configured": true, 00:31:49.281 "data_offset": 2048, 00:31:49.281 "data_size": 63488 00:31:49.281 }, 00:31:49.281 { 00:31:49.281 "name": "BaseBdev2", 00:31:49.281 "uuid": "949fd081-c278-59f1-9c08-d54bc625023f", 00:31:49.281 "is_configured": true, 00:31:49.281 "data_offset": 2048, 00:31:49.281 "data_size": 63488 00:31:49.281 }, 00:31:49.281 { 00:31:49.281 "name": "BaseBdev3", 00:31:49.281 "uuid": "59eda477-1d7f-5ac2-9bfb-d02bcbc4c870", 00:31:49.281 "is_configured": true, 00:31:49.281 "data_offset": 2048, 00:31:49.281 "data_size": 63488 00:31:49.281 }, 00:31:49.281 { 00:31:49.281 "name": "BaseBdev4", 00:31:49.281 "uuid": "d885740a-b18e-5b70-a7fd-311240bd474e", 00:31:49.281 "is_configured": true, 00:31:49.281 "data_offset": 2048, 00:31:49.281 "data_size": 63488 00:31:49.281 } 00:31:49.281 ] 00:31:49.281 }' 00:31:49.281 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:49.281 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:49.281 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:49.281 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:49.281 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:31:49.281 13:51:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:49.281 13:51:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:49.281 [2024-11-20 13:51:56.271657] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:31:49.539 [2024-11-20 13:51:56.360806] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:31:49.539 [2024-11-20 13:51:56.360903] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:49.539 [2024-11-20 13:51:56.360919] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:31:49.539 [2024-11-20 13:51:56.360928] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:31:49.539 13:51:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:49.539 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:31:49.539 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:49.539 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:49.539 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:49.539 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:49.539 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:31:49.539 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:49.539 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:49.539 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:49.539 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:49.539 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:49.539 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:49.540 13:51:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:49.540 13:51:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:49.540 13:51:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:49.540 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:49.540 "name": "raid_bdev1", 00:31:49.540 "uuid": "befab01c-3204-49aa-9025-c8f88189ffd6", 00:31:49.540 "strip_size_kb": 0, 00:31:49.540 "state": "online", 00:31:49.540 "raid_level": "raid1", 00:31:49.540 "superblock": true, 00:31:49.540 "num_base_bdevs": 4, 00:31:49.540 "num_base_bdevs_discovered": 3, 00:31:49.540 "num_base_bdevs_operational": 3, 00:31:49.540 "base_bdevs_list": [ 00:31:49.540 { 00:31:49.540 "name": null, 00:31:49.540 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:49.540 "is_configured": false, 00:31:49.540 "data_offset": 0, 00:31:49.540 "data_size": 63488 00:31:49.540 }, 00:31:49.540 { 00:31:49.540 "name": "BaseBdev2", 00:31:49.540 "uuid": "949fd081-c278-59f1-9c08-d54bc625023f", 00:31:49.540 "is_configured": true, 00:31:49.540 "data_offset": 2048, 00:31:49.540 "data_size": 63488 00:31:49.540 }, 00:31:49.540 { 00:31:49.540 "name": "BaseBdev3", 00:31:49.540 "uuid": "59eda477-1d7f-5ac2-9bfb-d02bcbc4c870", 00:31:49.540 "is_configured": true, 00:31:49.540 "data_offset": 2048, 00:31:49.540 "data_size": 63488 00:31:49.540 }, 00:31:49.540 { 00:31:49.540 "name": "BaseBdev4", 00:31:49.540 "uuid": "d885740a-b18e-5b70-a7fd-311240bd474e", 00:31:49.540 "is_configured": true, 00:31:49.540 "data_offset": 2048, 00:31:49.540 "data_size": 63488 00:31:49.540 } 00:31:49.540 ] 00:31:49.540 }' 00:31:49.540 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:49.540 13:51:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:49.797 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:31:49.797 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:49.797 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:31:49.797 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:31:49.797 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:49.797 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:49.797 13:51:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:49.797 13:51:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:49.797 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:49.797 13:51:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:49.797 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:49.797 "name": "raid_bdev1", 00:31:49.797 "uuid": "befab01c-3204-49aa-9025-c8f88189ffd6", 00:31:49.797 "strip_size_kb": 0, 00:31:49.797 "state": "online", 00:31:49.797 "raid_level": "raid1", 00:31:49.797 "superblock": true, 00:31:49.797 "num_base_bdevs": 4, 00:31:49.797 "num_base_bdevs_discovered": 3, 00:31:49.797 "num_base_bdevs_operational": 3, 00:31:49.798 "base_bdevs_list": [ 00:31:49.798 { 00:31:49.798 "name": null, 00:31:49.798 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:49.798 "is_configured": false, 00:31:49.798 "data_offset": 0, 00:31:49.798 "data_size": 63488 00:31:49.798 }, 00:31:49.798 { 00:31:49.798 "name": "BaseBdev2", 00:31:49.798 "uuid": "949fd081-c278-59f1-9c08-d54bc625023f", 00:31:49.798 "is_configured": true, 00:31:49.798 "data_offset": 2048, 00:31:49.798 "data_size": 63488 00:31:49.798 }, 00:31:49.798 { 00:31:49.798 "name": "BaseBdev3", 00:31:49.798 "uuid": "59eda477-1d7f-5ac2-9bfb-d02bcbc4c870", 00:31:49.798 "is_configured": true, 00:31:49.798 "data_offset": 2048, 00:31:49.798 "data_size": 63488 00:31:49.798 }, 00:31:49.798 { 00:31:49.798 "name": "BaseBdev4", 00:31:49.798 "uuid": "d885740a-b18e-5b70-a7fd-311240bd474e", 00:31:49.798 "is_configured": true, 00:31:49.798 "data_offset": 2048, 00:31:49.798 "data_size": 63488 00:31:49.798 } 00:31:49.798 ] 00:31:49.798 }' 00:31:49.798 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:49.798 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:31:49.798 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:49.798 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:31:49.798 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:31:49.798 13:51:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:49.798 13:51:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:49.798 [2024-11-20 13:51:56.797769] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:31:49.798 [2024-11-20 13:51:56.806124] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000ca35d0 00:31:49.798 13:51:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:49.798 13:51:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:31:49.798 [2024-11-20 13:51:56.807942] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:31:51.170 13:51:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:51.170 13:51:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:51.170 13:51:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:51.170 13:51:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:51.170 13:51:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:51.170 13:51:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:51.170 13:51:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:51.170 13:51:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:51.170 13:51:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:51.170 13:51:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:51.170 13:51:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:51.170 "name": "raid_bdev1", 00:31:51.170 "uuid": "befab01c-3204-49aa-9025-c8f88189ffd6", 00:31:51.170 "strip_size_kb": 0, 00:31:51.170 "state": "online", 00:31:51.170 "raid_level": "raid1", 00:31:51.170 "superblock": true, 00:31:51.170 "num_base_bdevs": 4, 00:31:51.170 "num_base_bdevs_discovered": 4, 00:31:51.170 "num_base_bdevs_operational": 4, 00:31:51.170 "process": { 00:31:51.170 "type": "rebuild", 00:31:51.170 "target": "spare", 00:31:51.170 "progress": { 00:31:51.170 "blocks": 20480, 00:31:51.170 "percent": 32 00:31:51.170 } 00:31:51.170 }, 00:31:51.170 "base_bdevs_list": [ 00:31:51.170 { 00:31:51.170 "name": "spare", 00:31:51.170 "uuid": "4c9ebeab-bb1e-5072-9065-732e55c929b5", 00:31:51.170 "is_configured": true, 00:31:51.170 "data_offset": 2048, 00:31:51.170 "data_size": 63488 00:31:51.170 }, 00:31:51.170 { 00:31:51.170 "name": "BaseBdev2", 00:31:51.170 "uuid": "949fd081-c278-59f1-9c08-d54bc625023f", 00:31:51.170 "is_configured": true, 00:31:51.170 "data_offset": 2048, 00:31:51.170 "data_size": 63488 00:31:51.170 }, 00:31:51.170 { 00:31:51.170 "name": "BaseBdev3", 00:31:51.170 "uuid": "59eda477-1d7f-5ac2-9bfb-d02bcbc4c870", 00:31:51.170 "is_configured": true, 00:31:51.170 "data_offset": 2048, 00:31:51.170 "data_size": 63488 00:31:51.170 }, 00:31:51.170 { 00:31:51.170 "name": "BaseBdev4", 00:31:51.170 "uuid": "d885740a-b18e-5b70-a7fd-311240bd474e", 00:31:51.170 "is_configured": true, 00:31:51.170 "data_offset": 2048, 00:31:51.170 "data_size": 63488 00:31:51.170 } 00:31:51.170 ] 00:31:51.170 }' 00:31:51.170 13:51:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:51.170 13:51:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:51.170 13:51:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:51.170 13:51:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:51.170 13:51:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:31:51.170 13:51:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:31:51.170 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:31:51.170 13:51:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:31:51.170 13:51:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:31:51.170 13:51:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:31:51.170 13:51:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:31:51.170 13:51:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:51.170 13:51:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:51.170 [2024-11-20 13:51:57.914097] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:31:51.170 [2024-11-20 13:51:58.014653] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000ca35d0 00:31:51.170 13:51:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:51.170 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:31:51.170 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:31:51.170 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:51.170 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:51.170 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:51.170 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:51.170 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:51.170 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:51.170 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:51.170 13:51:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:51.170 13:51:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:51.170 13:51:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:51.170 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:51.170 "name": "raid_bdev1", 00:31:51.170 "uuid": "befab01c-3204-49aa-9025-c8f88189ffd6", 00:31:51.170 "strip_size_kb": 0, 00:31:51.170 "state": "online", 00:31:51.170 "raid_level": "raid1", 00:31:51.170 "superblock": true, 00:31:51.170 "num_base_bdevs": 4, 00:31:51.170 "num_base_bdevs_discovered": 3, 00:31:51.170 "num_base_bdevs_operational": 3, 00:31:51.170 "process": { 00:31:51.170 "type": "rebuild", 00:31:51.170 "target": "spare", 00:31:51.170 "progress": { 00:31:51.170 "blocks": 22528, 00:31:51.170 "percent": 35 00:31:51.170 } 00:31:51.170 }, 00:31:51.170 "base_bdevs_list": [ 00:31:51.170 { 00:31:51.170 "name": "spare", 00:31:51.170 "uuid": "4c9ebeab-bb1e-5072-9065-732e55c929b5", 00:31:51.170 "is_configured": true, 00:31:51.170 "data_offset": 2048, 00:31:51.170 "data_size": 63488 00:31:51.170 }, 00:31:51.170 { 00:31:51.170 "name": null, 00:31:51.170 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:51.170 "is_configured": false, 00:31:51.170 "data_offset": 0, 00:31:51.170 "data_size": 63488 00:31:51.170 }, 00:31:51.170 { 00:31:51.170 "name": "BaseBdev3", 00:31:51.170 "uuid": "59eda477-1d7f-5ac2-9bfb-d02bcbc4c870", 00:31:51.170 "is_configured": true, 00:31:51.170 "data_offset": 2048, 00:31:51.171 "data_size": 63488 00:31:51.171 }, 00:31:51.171 { 00:31:51.171 "name": "BaseBdev4", 00:31:51.171 "uuid": "d885740a-b18e-5b70-a7fd-311240bd474e", 00:31:51.171 "is_configured": true, 00:31:51.171 "data_offset": 2048, 00:31:51.171 "data_size": 63488 00:31:51.171 } 00:31:51.171 ] 00:31:51.171 }' 00:31:51.171 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:51.171 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:51.171 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:51.171 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:51.171 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=365 00:31:51.171 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:51.171 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:51.171 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:51.171 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:51.171 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:51.171 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:51.171 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:51.171 13:51:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:51.171 13:51:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:51.171 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:51.171 13:51:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:51.171 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:51.171 "name": "raid_bdev1", 00:31:51.171 "uuid": "befab01c-3204-49aa-9025-c8f88189ffd6", 00:31:51.171 "strip_size_kb": 0, 00:31:51.171 "state": "online", 00:31:51.171 "raid_level": "raid1", 00:31:51.171 "superblock": true, 00:31:51.171 "num_base_bdevs": 4, 00:31:51.171 "num_base_bdevs_discovered": 3, 00:31:51.171 "num_base_bdevs_operational": 3, 00:31:51.171 "process": { 00:31:51.171 "type": "rebuild", 00:31:51.171 "target": "spare", 00:31:51.171 "progress": { 00:31:51.171 "blocks": 24576, 00:31:51.171 "percent": 38 00:31:51.171 } 00:31:51.171 }, 00:31:51.171 "base_bdevs_list": [ 00:31:51.171 { 00:31:51.171 "name": "spare", 00:31:51.171 "uuid": "4c9ebeab-bb1e-5072-9065-732e55c929b5", 00:31:51.171 "is_configured": true, 00:31:51.171 "data_offset": 2048, 00:31:51.171 "data_size": 63488 00:31:51.171 }, 00:31:51.171 { 00:31:51.171 "name": null, 00:31:51.171 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:51.171 "is_configured": false, 00:31:51.171 "data_offset": 0, 00:31:51.171 "data_size": 63488 00:31:51.171 }, 00:31:51.171 { 00:31:51.171 "name": "BaseBdev3", 00:31:51.171 "uuid": "59eda477-1d7f-5ac2-9bfb-d02bcbc4c870", 00:31:51.171 "is_configured": true, 00:31:51.171 "data_offset": 2048, 00:31:51.171 "data_size": 63488 00:31:51.171 }, 00:31:51.171 { 00:31:51.171 "name": "BaseBdev4", 00:31:51.171 "uuid": "d885740a-b18e-5b70-a7fd-311240bd474e", 00:31:51.171 "is_configured": true, 00:31:51.171 "data_offset": 2048, 00:31:51.171 "data_size": 63488 00:31:51.171 } 00:31:51.171 ] 00:31:51.171 }' 00:31:51.171 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:51.171 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:51.171 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:51.171 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:51.171 13:51:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:31:52.566 13:51:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:52.566 13:51:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:52.566 13:51:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:52.566 13:51:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:52.566 13:51:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:52.566 13:51:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:52.566 13:51:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:52.566 13:51:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:52.566 13:51:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:52.566 13:51:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:52.566 13:51:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:52.566 13:51:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:52.566 "name": "raid_bdev1", 00:31:52.566 "uuid": "befab01c-3204-49aa-9025-c8f88189ffd6", 00:31:52.566 "strip_size_kb": 0, 00:31:52.566 "state": "online", 00:31:52.566 "raid_level": "raid1", 00:31:52.566 "superblock": true, 00:31:52.566 "num_base_bdevs": 4, 00:31:52.566 "num_base_bdevs_discovered": 3, 00:31:52.566 "num_base_bdevs_operational": 3, 00:31:52.566 "process": { 00:31:52.566 "type": "rebuild", 00:31:52.566 "target": "spare", 00:31:52.566 "progress": { 00:31:52.566 "blocks": 47104, 00:31:52.566 "percent": 74 00:31:52.566 } 00:31:52.566 }, 00:31:52.566 "base_bdevs_list": [ 00:31:52.566 { 00:31:52.566 "name": "spare", 00:31:52.566 "uuid": "4c9ebeab-bb1e-5072-9065-732e55c929b5", 00:31:52.566 "is_configured": true, 00:31:52.566 "data_offset": 2048, 00:31:52.566 "data_size": 63488 00:31:52.566 }, 00:31:52.566 { 00:31:52.566 "name": null, 00:31:52.566 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:52.566 "is_configured": false, 00:31:52.566 "data_offset": 0, 00:31:52.566 "data_size": 63488 00:31:52.566 }, 00:31:52.566 { 00:31:52.566 "name": "BaseBdev3", 00:31:52.566 "uuid": "59eda477-1d7f-5ac2-9bfb-d02bcbc4c870", 00:31:52.566 "is_configured": true, 00:31:52.566 "data_offset": 2048, 00:31:52.566 "data_size": 63488 00:31:52.566 }, 00:31:52.566 { 00:31:52.566 "name": "BaseBdev4", 00:31:52.566 "uuid": "d885740a-b18e-5b70-a7fd-311240bd474e", 00:31:52.566 "is_configured": true, 00:31:52.566 "data_offset": 2048, 00:31:52.566 "data_size": 63488 00:31:52.566 } 00:31:52.566 ] 00:31:52.566 }' 00:31:52.566 13:51:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:52.566 13:51:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:52.566 13:51:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:52.566 13:51:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:52.566 13:51:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:31:53.187 [2024-11-20 13:52:00.026959] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:31:53.187 [2024-11-20 13:52:00.027236] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:31:53.188 [2024-11-20 13:52:00.027387] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:53.445 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:53.445 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:53.445 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:53.445 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:53.446 "name": "raid_bdev1", 00:31:53.446 "uuid": "befab01c-3204-49aa-9025-c8f88189ffd6", 00:31:53.446 "strip_size_kb": 0, 00:31:53.446 "state": "online", 00:31:53.446 "raid_level": "raid1", 00:31:53.446 "superblock": true, 00:31:53.446 "num_base_bdevs": 4, 00:31:53.446 "num_base_bdevs_discovered": 3, 00:31:53.446 "num_base_bdevs_operational": 3, 00:31:53.446 "base_bdevs_list": [ 00:31:53.446 { 00:31:53.446 "name": "spare", 00:31:53.446 "uuid": "4c9ebeab-bb1e-5072-9065-732e55c929b5", 00:31:53.446 "is_configured": true, 00:31:53.446 "data_offset": 2048, 00:31:53.446 "data_size": 63488 00:31:53.446 }, 00:31:53.446 { 00:31:53.446 "name": null, 00:31:53.446 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:53.446 "is_configured": false, 00:31:53.446 "data_offset": 0, 00:31:53.446 "data_size": 63488 00:31:53.446 }, 00:31:53.446 { 00:31:53.446 "name": "BaseBdev3", 00:31:53.446 "uuid": "59eda477-1d7f-5ac2-9bfb-d02bcbc4c870", 00:31:53.446 "is_configured": true, 00:31:53.446 "data_offset": 2048, 00:31:53.446 "data_size": 63488 00:31:53.446 }, 00:31:53.446 { 00:31:53.446 "name": "BaseBdev4", 00:31:53.446 "uuid": "d885740a-b18e-5b70-a7fd-311240bd474e", 00:31:53.446 "is_configured": true, 00:31:53.446 "data_offset": 2048, 00:31:53.446 "data_size": 63488 00:31:53.446 } 00:31:53.446 ] 00:31:53.446 }' 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:53.446 "name": "raid_bdev1", 00:31:53.446 "uuid": "befab01c-3204-49aa-9025-c8f88189ffd6", 00:31:53.446 "strip_size_kb": 0, 00:31:53.446 "state": "online", 00:31:53.446 "raid_level": "raid1", 00:31:53.446 "superblock": true, 00:31:53.446 "num_base_bdevs": 4, 00:31:53.446 "num_base_bdevs_discovered": 3, 00:31:53.446 "num_base_bdevs_operational": 3, 00:31:53.446 "base_bdevs_list": [ 00:31:53.446 { 00:31:53.446 "name": "spare", 00:31:53.446 "uuid": "4c9ebeab-bb1e-5072-9065-732e55c929b5", 00:31:53.446 "is_configured": true, 00:31:53.446 "data_offset": 2048, 00:31:53.446 "data_size": 63488 00:31:53.446 }, 00:31:53.446 { 00:31:53.446 "name": null, 00:31:53.446 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:53.446 "is_configured": false, 00:31:53.446 "data_offset": 0, 00:31:53.446 "data_size": 63488 00:31:53.446 }, 00:31:53.446 { 00:31:53.446 "name": "BaseBdev3", 00:31:53.446 "uuid": "59eda477-1d7f-5ac2-9bfb-d02bcbc4c870", 00:31:53.446 "is_configured": true, 00:31:53.446 "data_offset": 2048, 00:31:53.446 "data_size": 63488 00:31:53.446 }, 00:31:53.446 { 00:31:53.446 "name": "BaseBdev4", 00:31:53.446 "uuid": "d885740a-b18e-5b70-a7fd-311240bd474e", 00:31:53.446 "is_configured": true, 00:31:53.446 "data_offset": 2048, 00:31:53.446 "data_size": 63488 00:31:53.446 } 00:31:53.446 ] 00:31:53.446 }' 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:31:53.446 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:53.705 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:31:53.705 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:31:53.705 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:53.705 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:53.705 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:53.705 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:53.705 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:31:53.705 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:53.705 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:53.705 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:53.705 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:53.705 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:53.705 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:53.705 13:52:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:53.705 13:52:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:53.705 13:52:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:53.705 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:53.705 "name": "raid_bdev1", 00:31:53.705 "uuid": "befab01c-3204-49aa-9025-c8f88189ffd6", 00:31:53.705 "strip_size_kb": 0, 00:31:53.705 "state": "online", 00:31:53.705 "raid_level": "raid1", 00:31:53.705 "superblock": true, 00:31:53.705 "num_base_bdevs": 4, 00:31:53.705 "num_base_bdevs_discovered": 3, 00:31:53.705 "num_base_bdevs_operational": 3, 00:31:53.705 "base_bdevs_list": [ 00:31:53.705 { 00:31:53.705 "name": "spare", 00:31:53.705 "uuid": "4c9ebeab-bb1e-5072-9065-732e55c929b5", 00:31:53.705 "is_configured": true, 00:31:53.705 "data_offset": 2048, 00:31:53.705 "data_size": 63488 00:31:53.705 }, 00:31:53.705 { 00:31:53.705 "name": null, 00:31:53.705 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:53.705 "is_configured": false, 00:31:53.705 "data_offset": 0, 00:31:53.705 "data_size": 63488 00:31:53.705 }, 00:31:53.705 { 00:31:53.705 "name": "BaseBdev3", 00:31:53.705 "uuid": "59eda477-1d7f-5ac2-9bfb-d02bcbc4c870", 00:31:53.705 "is_configured": true, 00:31:53.705 "data_offset": 2048, 00:31:53.705 "data_size": 63488 00:31:53.705 }, 00:31:53.705 { 00:31:53.705 "name": "BaseBdev4", 00:31:53.705 "uuid": "d885740a-b18e-5b70-a7fd-311240bd474e", 00:31:53.705 "is_configured": true, 00:31:53.705 "data_offset": 2048, 00:31:53.705 "data_size": 63488 00:31:53.705 } 00:31:53.705 ] 00:31:53.705 }' 00:31:53.705 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:53.705 13:52:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:53.963 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:31:53.963 13:52:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:53.963 13:52:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:53.963 [2024-11-20 13:52:00.832948] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:31:53.963 [2024-11-20 13:52:00.833005] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:31:53.963 [2024-11-20 13:52:00.833093] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:31:53.963 [2024-11-20 13:52:00.833169] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:31:53.963 [2024-11-20 13:52:00.833179] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:31:53.963 13:52:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:53.963 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:53.963 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:31:53.963 13:52:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:53.963 13:52:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:53.963 13:52:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:53.963 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:31:53.963 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:31:53.963 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:31:53.963 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:31:53.963 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:31:53.963 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:31:53.963 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:31:53.963 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:31:53.963 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:31:53.963 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:31:53.963 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:31:53.963 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:31:53.963 13:52:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:31:54.222 /dev/nbd0 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:31:54.222 1+0 records in 00:31:54.222 1+0 records out 00:31:54.222 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000227563 s, 18.0 MB/s 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:31:54.222 /dev/nbd1 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:31:54.222 1+0 records in 00:31:54.222 1+0 records out 00:31:54.222 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000325119 s, 12.6 MB/s 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:31:54.222 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:31:54.480 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:31:54.480 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:31:54.480 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:31:54.480 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:31:54.480 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:31:54.480 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:31:54.480 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:31:54.738 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:31:54.738 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:31:54.738 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:31:54.738 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:31:54.738 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:31:54.738 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:31:54.738 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:31:54.738 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:31:54.738 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:31:54.738 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:54.997 [2024-11-20 13:52:01.853655] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:31:54.997 [2024-11-20 13:52:01.853799] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:54.997 [2024-11-20 13:52:01.853828] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:31:54.997 [2024-11-20 13:52:01.853836] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:54.997 [2024-11-20 13:52:01.855896] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:54.997 [2024-11-20 13:52:01.855927] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:31:54.997 [2024-11-20 13:52:01.856027] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:31:54.997 [2024-11-20 13:52:01.856071] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:31:54.997 [2024-11-20 13:52:01.856187] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:31:54.997 [2024-11-20 13:52:01.856262] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:31:54.997 spare 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:54.997 [2024-11-20 13:52:01.956344] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:31:54.997 [2024-11-20 13:52:01.956478] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:31:54.997 [2024-11-20 13:52:01.956793] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1c80 00:31:54.997 [2024-11-20 13:52:01.956966] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:31:54.997 [2024-11-20 13:52:01.956975] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:31:54.997 [2024-11-20 13:52:01.957146] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:54.997 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:54.997 "name": "raid_bdev1", 00:31:54.997 "uuid": "befab01c-3204-49aa-9025-c8f88189ffd6", 00:31:54.997 "strip_size_kb": 0, 00:31:54.997 "state": "online", 00:31:54.997 "raid_level": "raid1", 00:31:54.997 "superblock": true, 00:31:54.997 "num_base_bdevs": 4, 00:31:54.997 "num_base_bdevs_discovered": 3, 00:31:54.997 "num_base_bdevs_operational": 3, 00:31:54.997 "base_bdevs_list": [ 00:31:54.997 { 00:31:54.997 "name": "spare", 00:31:54.997 "uuid": "4c9ebeab-bb1e-5072-9065-732e55c929b5", 00:31:54.997 "is_configured": true, 00:31:54.998 "data_offset": 2048, 00:31:54.998 "data_size": 63488 00:31:54.998 }, 00:31:54.998 { 00:31:54.998 "name": null, 00:31:54.998 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:54.998 "is_configured": false, 00:31:54.998 "data_offset": 2048, 00:31:54.998 "data_size": 63488 00:31:54.998 }, 00:31:54.998 { 00:31:54.998 "name": "BaseBdev3", 00:31:54.998 "uuid": "59eda477-1d7f-5ac2-9bfb-d02bcbc4c870", 00:31:54.998 "is_configured": true, 00:31:54.998 "data_offset": 2048, 00:31:54.998 "data_size": 63488 00:31:54.998 }, 00:31:54.998 { 00:31:54.998 "name": "BaseBdev4", 00:31:54.998 "uuid": "d885740a-b18e-5b70-a7fd-311240bd474e", 00:31:54.998 "is_configured": true, 00:31:54.998 "data_offset": 2048, 00:31:54.998 "data_size": 63488 00:31:54.998 } 00:31:54.998 ] 00:31:54.998 }' 00:31:54.998 13:52:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:54.998 13:52:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:55.255 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:31:55.255 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:55.255 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:31:55.255 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:31:55.255 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:55.255 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:55.255 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:55.255 13:52:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:55.255 13:52:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:55.255 13:52:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:55.256 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:55.256 "name": "raid_bdev1", 00:31:55.256 "uuid": "befab01c-3204-49aa-9025-c8f88189ffd6", 00:31:55.256 "strip_size_kb": 0, 00:31:55.256 "state": "online", 00:31:55.256 "raid_level": "raid1", 00:31:55.256 "superblock": true, 00:31:55.256 "num_base_bdevs": 4, 00:31:55.256 "num_base_bdevs_discovered": 3, 00:31:55.256 "num_base_bdevs_operational": 3, 00:31:55.256 "base_bdevs_list": [ 00:31:55.256 { 00:31:55.256 "name": "spare", 00:31:55.256 "uuid": "4c9ebeab-bb1e-5072-9065-732e55c929b5", 00:31:55.256 "is_configured": true, 00:31:55.256 "data_offset": 2048, 00:31:55.256 "data_size": 63488 00:31:55.256 }, 00:31:55.256 { 00:31:55.256 "name": null, 00:31:55.256 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:55.256 "is_configured": false, 00:31:55.256 "data_offset": 2048, 00:31:55.256 "data_size": 63488 00:31:55.256 }, 00:31:55.256 { 00:31:55.256 "name": "BaseBdev3", 00:31:55.256 "uuid": "59eda477-1d7f-5ac2-9bfb-d02bcbc4c870", 00:31:55.256 "is_configured": true, 00:31:55.256 "data_offset": 2048, 00:31:55.256 "data_size": 63488 00:31:55.256 }, 00:31:55.256 { 00:31:55.256 "name": "BaseBdev4", 00:31:55.256 "uuid": "d885740a-b18e-5b70-a7fd-311240bd474e", 00:31:55.256 "is_configured": true, 00:31:55.256 "data_offset": 2048, 00:31:55.256 "data_size": 63488 00:31:55.256 } 00:31:55.256 ] 00:31:55.256 }' 00:31:55.256 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:55.513 [2024-11-20 13:52:02.401817] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:55.513 "name": "raid_bdev1", 00:31:55.513 "uuid": "befab01c-3204-49aa-9025-c8f88189ffd6", 00:31:55.513 "strip_size_kb": 0, 00:31:55.513 "state": "online", 00:31:55.513 "raid_level": "raid1", 00:31:55.513 "superblock": true, 00:31:55.513 "num_base_bdevs": 4, 00:31:55.513 "num_base_bdevs_discovered": 2, 00:31:55.513 "num_base_bdevs_operational": 2, 00:31:55.513 "base_bdevs_list": [ 00:31:55.513 { 00:31:55.513 "name": null, 00:31:55.513 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:55.513 "is_configured": false, 00:31:55.513 "data_offset": 0, 00:31:55.513 "data_size": 63488 00:31:55.513 }, 00:31:55.513 { 00:31:55.513 "name": null, 00:31:55.513 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:55.513 "is_configured": false, 00:31:55.513 "data_offset": 2048, 00:31:55.513 "data_size": 63488 00:31:55.513 }, 00:31:55.513 { 00:31:55.513 "name": "BaseBdev3", 00:31:55.513 "uuid": "59eda477-1d7f-5ac2-9bfb-d02bcbc4c870", 00:31:55.513 "is_configured": true, 00:31:55.513 "data_offset": 2048, 00:31:55.513 "data_size": 63488 00:31:55.513 }, 00:31:55.513 { 00:31:55.513 "name": "BaseBdev4", 00:31:55.513 "uuid": "d885740a-b18e-5b70-a7fd-311240bd474e", 00:31:55.513 "is_configured": true, 00:31:55.513 "data_offset": 2048, 00:31:55.513 "data_size": 63488 00:31:55.513 } 00:31:55.513 ] 00:31:55.513 }' 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:55.513 13:52:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:55.771 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:31:55.771 13:52:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:55.771 13:52:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:55.771 [2024-11-20 13:52:02.789901] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:31:55.771 [2024-11-20 13:52:02.790106] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:31:55.771 [2024-11-20 13:52:02.790119] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:31:55.771 [2024-11-20 13:52:02.790154] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:31:55.771 [2024-11-20 13:52:02.797937] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1d50 00:31:55.771 13:52:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:55.771 13:52:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:31:55.771 [2024-11-20 13:52:02.799702] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:31:57.144 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:57.144 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:57.144 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:57.144 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:57.144 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:57.144 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:57.144 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:57.144 13:52:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:57.144 13:52:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:57.144 13:52:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:57.144 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:57.144 "name": "raid_bdev1", 00:31:57.144 "uuid": "befab01c-3204-49aa-9025-c8f88189ffd6", 00:31:57.144 "strip_size_kb": 0, 00:31:57.144 "state": "online", 00:31:57.144 "raid_level": "raid1", 00:31:57.144 "superblock": true, 00:31:57.144 "num_base_bdevs": 4, 00:31:57.144 "num_base_bdevs_discovered": 3, 00:31:57.144 "num_base_bdevs_operational": 3, 00:31:57.144 "process": { 00:31:57.144 "type": "rebuild", 00:31:57.144 "target": "spare", 00:31:57.144 "progress": { 00:31:57.144 "blocks": 20480, 00:31:57.144 "percent": 32 00:31:57.144 } 00:31:57.144 }, 00:31:57.144 "base_bdevs_list": [ 00:31:57.144 { 00:31:57.144 "name": "spare", 00:31:57.144 "uuid": "4c9ebeab-bb1e-5072-9065-732e55c929b5", 00:31:57.144 "is_configured": true, 00:31:57.144 "data_offset": 2048, 00:31:57.144 "data_size": 63488 00:31:57.144 }, 00:31:57.144 { 00:31:57.144 "name": null, 00:31:57.144 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:57.144 "is_configured": false, 00:31:57.144 "data_offset": 2048, 00:31:57.144 "data_size": 63488 00:31:57.144 }, 00:31:57.144 { 00:31:57.144 "name": "BaseBdev3", 00:31:57.144 "uuid": "59eda477-1d7f-5ac2-9bfb-d02bcbc4c870", 00:31:57.144 "is_configured": true, 00:31:57.144 "data_offset": 2048, 00:31:57.144 "data_size": 63488 00:31:57.144 }, 00:31:57.144 { 00:31:57.144 "name": "BaseBdev4", 00:31:57.144 "uuid": "d885740a-b18e-5b70-a7fd-311240bd474e", 00:31:57.144 "is_configured": true, 00:31:57.144 "data_offset": 2048, 00:31:57.144 "data_size": 63488 00:31:57.144 } 00:31:57.144 ] 00:31:57.144 }' 00:31:57.144 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:57.144 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:57.144 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:57.144 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:57.144 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:31:57.144 13:52:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:57.144 13:52:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:57.144 [2024-11-20 13:52:03.897693] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:31:57.144 [2024-11-20 13:52:03.906883] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:31:57.144 [2024-11-20 13:52:03.906932] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:57.144 [2024-11-20 13:52:03.906948] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:31:57.144 [2024-11-20 13:52:03.906954] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:31:57.144 13:52:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:57.144 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:31:57.144 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:57.144 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:57.144 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:57.145 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:57.145 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:31:57.145 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:57.145 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:57.145 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:57.145 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:57.145 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:57.145 13:52:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:57.145 13:52:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:57.145 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:57.145 13:52:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:57.145 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:57.145 "name": "raid_bdev1", 00:31:57.145 "uuid": "befab01c-3204-49aa-9025-c8f88189ffd6", 00:31:57.145 "strip_size_kb": 0, 00:31:57.145 "state": "online", 00:31:57.145 "raid_level": "raid1", 00:31:57.145 "superblock": true, 00:31:57.145 "num_base_bdevs": 4, 00:31:57.145 "num_base_bdevs_discovered": 2, 00:31:57.145 "num_base_bdevs_operational": 2, 00:31:57.145 "base_bdevs_list": [ 00:31:57.145 { 00:31:57.145 "name": null, 00:31:57.145 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:57.145 "is_configured": false, 00:31:57.145 "data_offset": 0, 00:31:57.145 "data_size": 63488 00:31:57.145 }, 00:31:57.145 { 00:31:57.145 "name": null, 00:31:57.145 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:57.145 "is_configured": false, 00:31:57.145 "data_offset": 2048, 00:31:57.145 "data_size": 63488 00:31:57.145 }, 00:31:57.145 { 00:31:57.145 "name": "BaseBdev3", 00:31:57.145 "uuid": "59eda477-1d7f-5ac2-9bfb-d02bcbc4c870", 00:31:57.145 "is_configured": true, 00:31:57.145 "data_offset": 2048, 00:31:57.145 "data_size": 63488 00:31:57.145 }, 00:31:57.145 { 00:31:57.145 "name": "BaseBdev4", 00:31:57.145 "uuid": "d885740a-b18e-5b70-a7fd-311240bd474e", 00:31:57.145 "is_configured": true, 00:31:57.145 "data_offset": 2048, 00:31:57.145 "data_size": 63488 00:31:57.145 } 00:31:57.145 ] 00:31:57.145 }' 00:31:57.145 13:52:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:57.145 13:52:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:57.402 13:52:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:31:57.402 13:52:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:57.402 13:52:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:57.402 [2024-11-20 13:52:04.236465] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:31:57.402 [2024-11-20 13:52:04.236534] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:57.402 [2024-11-20 13:52:04.236565] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c380 00:31:57.402 [2024-11-20 13:52:04.236574] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:57.402 [2024-11-20 13:52:04.237048] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:57.402 [2024-11-20 13:52:04.237159] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:31:57.402 [2024-11-20 13:52:04.237274] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:31:57.403 [2024-11-20 13:52:04.237286] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:31:57.403 [2024-11-20 13:52:04.237304] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:31:57.403 [2024-11-20 13:52:04.237329] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:31:57.403 [2024-11-20 13:52:04.245339] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1e20 00:31:57.403 spare 00:31:57.403 13:52:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:57.403 13:52:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:31:57.403 [2024-11-20 13:52:04.247178] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:31:58.336 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:58.336 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:58.336 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:58.336 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:58.336 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:58.336 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:58.336 13:52:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:58.336 13:52:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:58.336 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:58.336 13:52:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:58.336 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:58.336 "name": "raid_bdev1", 00:31:58.336 "uuid": "befab01c-3204-49aa-9025-c8f88189ffd6", 00:31:58.336 "strip_size_kb": 0, 00:31:58.336 "state": "online", 00:31:58.336 "raid_level": "raid1", 00:31:58.336 "superblock": true, 00:31:58.336 "num_base_bdevs": 4, 00:31:58.336 "num_base_bdevs_discovered": 3, 00:31:58.336 "num_base_bdevs_operational": 3, 00:31:58.336 "process": { 00:31:58.336 "type": "rebuild", 00:31:58.336 "target": "spare", 00:31:58.336 "progress": { 00:31:58.336 "blocks": 20480, 00:31:58.336 "percent": 32 00:31:58.336 } 00:31:58.336 }, 00:31:58.336 "base_bdevs_list": [ 00:31:58.336 { 00:31:58.336 "name": "spare", 00:31:58.336 "uuid": "4c9ebeab-bb1e-5072-9065-732e55c929b5", 00:31:58.336 "is_configured": true, 00:31:58.336 "data_offset": 2048, 00:31:58.336 "data_size": 63488 00:31:58.336 }, 00:31:58.336 { 00:31:58.336 "name": null, 00:31:58.336 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:58.336 "is_configured": false, 00:31:58.336 "data_offset": 2048, 00:31:58.336 "data_size": 63488 00:31:58.336 }, 00:31:58.336 { 00:31:58.336 "name": "BaseBdev3", 00:31:58.336 "uuid": "59eda477-1d7f-5ac2-9bfb-d02bcbc4c870", 00:31:58.336 "is_configured": true, 00:31:58.336 "data_offset": 2048, 00:31:58.336 "data_size": 63488 00:31:58.336 }, 00:31:58.336 { 00:31:58.336 "name": "BaseBdev4", 00:31:58.336 "uuid": "d885740a-b18e-5b70-a7fd-311240bd474e", 00:31:58.336 "is_configured": true, 00:31:58.336 "data_offset": 2048, 00:31:58.336 "data_size": 63488 00:31:58.336 } 00:31:58.336 ] 00:31:58.336 }' 00:31:58.336 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:58.336 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:58.336 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:58.336 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:58.336 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:31:58.336 13:52:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:58.336 13:52:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:58.336 [2024-11-20 13:52:05.356895] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:31:58.595 [2024-11-20 13:52:05.455240] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:31:58.595 [2024-11-20 13:52:05.455375] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:58.595 [2024-11-20 13:52:05.455391] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:31:58.595 [2024-11-20 13:52:05.455399] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:31:58.595 13:52:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:58.595 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:31:58.595 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:58.595 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:58.595 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:58.595 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:58.595 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:31:58.595 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:58.595 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:58.595 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:58.595 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:58.595 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:58.595 13:52:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:58.595 13:52:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:58.595 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:58.595 13:52:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:58.595 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:58.595 "name": "raid_bdev1", 00:31:58.595 "uuid": "befab01c-3204-49aa-9025-c8f88189ffd6", 00:31:58.595 "strip_size_kb": 0, 00:31:58.595 "state": "online", 00:31:58.595 "raid_level": "raid1", 00:31:58.595 "superblock": true, 00:31:58.595 "num_base_bdevs": 4, 00:31:58.595 "num_base_bdevs_discovered": 2, 00:31:58.595 "num_base_bdevs_operational": 2, 00:31:58.595 "base_bdevs_list": [ 00:31:58.595 { 00:31:58.595 "name": null, 00:31:58.595 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:58.595 "is_configured": false, 00:31:58.595 "data_offset": 0, 00:31:58.595 "data_size": 63488 00:31:58.595 }, 00:31:58.596 { 00:31:58.596 "name": null, 00:31:58.596 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:58.596 "is_configured": false, 00:31:58.596 "data_offset": 2048, 00:31:58.596 "data_size": 63488 00:31:58.596 }, 00:31:58.596 { 00:31:58.596 "name": "BaseBdev3", 00:31:58.596 "uuid": "59eda477-1d7f-5ac2-9bfb-d02bcbc4c870", 00:31:58.596 "is_configured": true, 00:31:58.596 "data_offset": 2048, 00:31:58.596 "data_size": 63488 00:31:58.596 }, 00:31:58.596 { 00:31:58.596 "name": "BaseBdev4", 00:31:58.596 "uuid": "d885740a-b18e-5b70-a7fd-311240bd474e", 00:31:58.596 "is_configured": true, 00:31:58.596 "data_offset": 2048, 00:31:58.596 "data_size": 63488 00:31:58.596 } 00:31:58.596 ] 00:31:58.596 }' 00:31:58.596 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:58.596 13:52:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:58.908 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:31:58.908 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:58.908 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:31:58.908 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:31:58.908 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:58.908 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:58.908 13:52:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:58.908 13:52:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:58.908 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:58.908 13:52:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:58.908 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:58.908 "name": "raid_bdev1", 00:31:58.908 "uuid": "befab01c-3204-49aa-9025-c8f88189ffd6", 00:31:58.908 "strip_size_kb": 0, 00:31:58.908 "state": "online", 00:31:58.908 "raid_level": "raid1", 00:31:58.908 "superblock": true, 00:31:58.908 "num_base_bdevs": 4, 00:31:58.908 "num_base_bdevs_discovered": 2, 00:31:58.908 "num_base_bdevs_operational": 2, 00:31:58.908 "base_bdevs_list": [ 00:31:58.908 { 00:31:58.908 "name": null, 00:31:58.908 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:58.908 "is_configured": false, 00:31:58.908 "data_offset": 0, 00:31:58.908 "data_size": 63488 00:31:58.908 }, 00:31:58.908 { 00:31:58.908 "name": null, 00:31:58.908 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:58.908 "is_configured": false, 00:31:58.908 "data_offset": 2048, 00:31:58.908 "data_size": 63488 00:31:58.908 }, 00:31:58.908 { 00:31:58.908 "name": "BaseBdev3", 00:31:58.908 "uuid": "59eda477-1d7f-5ac2-9bfb-d02bcbc4c870", 00:31:58.908 "is_configured": true, 00:31:58.908 "data_offset": 2048, 00:31:58.908 "data_size": 63488 00:31:58.908 }, 00:31:58.908 { 00:31:58.908 "name": "BaseBdev4", 00:31:58.908 "uuid": "d885740a-b18e-5b70-a7fd-311240bd474e", 00:31:58.908 "is_configured": true, 00:31:58.908 "data_offset": 2048, 00:31:58.908 "data_size": 63488 00:31:58.908 } 00:31:58.908 ] 00:31:58.908 }' 00:31:58.908 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:58.908 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:31:58.908 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:58.908 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:31:58.908 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:31:58.908 13:52:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:58.908 13:52:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:58.908 13:52:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:58.908 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:31:58.908 13:52:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:58.908 13:52:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:58.908 [2024-11-20 13:52:05.884484] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:31:58.908 [2024-11-20 13:52:05.884545] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:58.908 [2024-11-20 13:52:05.884565] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c980 00:31:58.908 [2024-11-20 13:52:05.884574] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:58.908 [2024-11-20 13:52:05.885014] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:58.908 [2024-11-20 13:52:05.885034] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:31:58.908 [2024-11-20 13:52:05.885105] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:31:58.908 [2024-11-20 13:52:05.885120] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:31:58.908 [2024-11-20 13:52:05.885127] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:31:58.908 [2024-11-20 13:52:05.885141] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:31:58.908 BaseBdev1 00:31:58.908 13:52:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:58.908 13:52:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:31:59.841 13:52:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:31:59.841 13:52:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:59.841 13:52:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:59.841 13:52:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:59.841 13:52:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:59.841 13:52:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:31:59.841 13:52:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:59.841 13:52:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:59.842 13:52:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:59.842 13:52:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:59.842 13:52:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:59.842 13:52:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:59.842 13:52:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:59.842 13:52:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:00.099 13:52:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:00.099 13:52:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:00.099 "name": "raid_bdev1", 00:32:00.099 "uuid": "befab01c-3204-49aa-9025-c8f88189ffd6", 00:32:00.099 "strip_size_kb": 0, 00:32:00.099 "state": "online", 00:32:00.099 "raid_level": "raid1", 00:32:00.099 "superblock": true, 00:32:00.099 "num_base_bdevs": 4, 00:32:00.099 "num_base_bdevs_discovered": 2, 00:32:00.099 "num_base_bdevs_operational": 2, 00:32:00.099 "base_bdevs_list": [ 00:32:00.099 { 00:32:00.099 "name": null, 00:32:00.099 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:00.099 "is_configured": false, 00:32:00.099 "data_offset": 0, 00:32:00.099 "data_size": 63488 00:32:00.099 }, 00:32:00.099 { 00:32:00.099 "name": null, 00:32:00.099 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:00.099 "is_configured": false, 00:32:00.099 "data_offset": 2048, 00:32:00.099 "data_size": 63488 00:32:00.099 }, 00:32:00.099 { 00:32:00.099 "name": "BaseBdev3", 00:32:00.099 "uuid": "59eda477-1d7f-5ac2-9bfb-d02bcbc4c870", 00:32:00.099 "is_configured": true, 00:32:00.099 "data_offset": 2048, 00:32:00.099 "data_size": 63488 00:32:00.099 }, 00:32:00.099 { 00:32:00.099 "name": "BaseBdev4", 00:32:00.099 "uuid": "d885740a-b18e-5b70-a7fd-311240bd474e", 00:32:00.099 "is_configured": true, 00:32:00.099 "data_offset": 2048, 00:32:00.099 "data_size": 63488 00:32:00.099 } 00:32:00.099 ] 00:32:00.099 }' 00:32:00.099 13:52:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:00.099 13:52:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:00.358 13:52:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:32:00.358 13:52:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:00.358 13:52:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:32:00.358 13:52:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:32:00.358 13:52:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:00.358 13:52:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:00.358 13:52:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:00.358 13:52:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:00.358 13:52:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:00.358 13:52:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:00.358 13:52:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:00.358 "name": "raid_bdev1", 00:32:00.358 "uuid": "befab01c-3204-49aa-9025-c8f88189ffd6", 00:32:00.358 "strip_size_kb": 0, 00:32:00.358 "state": "online", 00:32:00.358 "raid_level": "raid1", 00:32:00.358 "superblock": true, 00:32:00.358 "num_base_bdevs": 4, 00:32:00.358 "num_base_bdevs_discovered": 2, 00:32:00.358 "num_base_bdevs_operational": 2, 00:32:00.358 "base_bdevs_list": [ 00:32:00.358 { 00:32:00.358 "name": null, 00:32:00.358 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:00.358 "is_configured": false, 00:32:00.358 "data_offset": 0, 00:32:00.358 "data_size": 63488 00:32:00.358 }, 00:32:00.358 { 00:32:00.358 "name": null, 00:32:00.358 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:00.358 "is_configured": false, 00:32:00.358 "data_offset": 2048, 00:32:00.358 "data_size": 63488 00:32:00.358 }, 00:32:00.358 { 00:32:00.358 "name": "BaseBdev3", 00:32:00.358 "uuid": "59eda477-1d7f-5ac2-9bfb-d02bcbc4c870", 00:32:00.358 "is_configured": true, 00:32:00.358 "data_offset": 2048, 00:32:00.358 "data_size": 63488 00:32:00.358 }, 00:32:00.358 { 00:32:00.358 "name": "BaseBdev4", 00:32:00.358 "uuid": "d885740a-b18e-5b70-a7fd-311240bd474e", 00:32:00.358 "is_configured": true, 00:32:00.358 "data_offset": 2048, 00:32:00.358 "data_size": 63488 00:32:00.358 } 00:32:00.358 ] 00:32:00.358 }' 00:32:00.358 13:52:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:00.359 13:52:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:32:00.359 13:52:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:00.359 13:52:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:32:00.359 13:52:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:32:00.359 13:52:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:32:00.359 13:52:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:32:00.359 13:52:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:32:00.359 13:52:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:32:00.359 13:52:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:32:00.359 13:52:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:32:00.359 13:52:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:32:00.359 13:52:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:00.359 13:52:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:00.359 [2024-11-20 13:52:07.332798] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:32:00.359 [2024-11-20 13:52:07.333012] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:32:00.359 [2024-11-20 13:52:07.333026] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:32:00.359 request: 00:32:00.359 { 00:32:00.359 "base_bdev": "BaseBdev1", 00:32:00.359 "raid_bdev": "raid_bdev1", 00:32:00.359 "method": "bdev_raid_add_base_bdev", 00:32:00.359 "req_id": 1 00:32:00.359 } 00:32:00.359 Got JSON-RPC error response 00:32:00.359 response: 00:32:00.359 { 00:32:00.359 "code": -22, 00:32:00.359 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:32:00.359 } 00:32:00.359 13:52:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:32:00.359 13:52:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:32:00.359 13:52:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:32:00.359 13:52:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:32:00.359 13:52:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:32:00.359 13:52:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:32:01.293 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:32:01.293 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:01.293 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:01.293 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:01.293 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:01.293 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:01.293 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:01.293 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:01.293 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:01.293 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:01.293 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:01.293 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:01.293 13:52:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:01.293 13:52:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:01.609 13:52:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:01.609 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:01.609 "name": "raid_bdev1", 00:32:01.609 "uuid": "befab01c-3204-49aa-9025-c8f88189ffd6", 00:32:01.609 "strip_size_kb": 0, 00:32:01.609 "state": "online", 00:32:01.609 "raid_level": "raid1", 00:32:01.609 "superblock": true, 00:32:01.609 "num_base_bdevs": 4, 00:32:01.609 "num_base_bdevs_discovered": 2, 00:32:01.609 "num_base_bdevs_operational": 2, 00:32:01.609 "base_bdevs_list": [ 00:32:01.609 { 00:32:01.609 "name": null, 00:32:01.609 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:01.609 "is_configured": false, 00:32:01.609 "data_offset": 0, 00:32:01.609 "data_size": 63488 00:32:01.609 }, 00:32:01.609 { 00:32:01.609 "name": null, 00:32:01.609 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:01.609 "is_configured": false, 00:32:01.610 "data_offset": 2048, 00:32:01.610 "data_size": 63488 00:32:01.610 }, 00:32:01.610 { 00:32:01.610 "name": "BaseBdev3", 00:32:01.610 "uuid": "59eda477-1d7f-5ac2-9bfb-d02bcbc4c870", 00:32:01.610 "is_configured": true, 00:32:01.610 "data_offset": 2048, 00:32:01.610 "data_size": 63488 00:32:01.610 }, 00:32:01.610 { 00:32:01.610 "name": "BaseBdev4", 00:32:01.610 "uuid": "d885740a-b18e-5b70-a7fd-311240bd474e", 00:32:01.610 "is_configured": true, 00:32:01.610 "data_offset": 2048, 00:32:01.610 "data_size": 63488 00:32:01.610 } 00:32:01.610 ] 00:32:01.610 }' 00:32:01.610 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:01.610 13:52:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:01.891 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:32:01.891 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:01.891 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:32:01.891 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:32:01.891 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:01.891 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:01.891 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:01.891 13:52:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:01.891 13:52:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:01.891 13:52:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:01.891 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:01.891 "name": "raid_bdev1", 00:32:01.891 "uuid": "befab01c-3204-49aa-9025-c8f88189ffd6", 00:32:01.891 "strip_size_kb": 0, 00:32:01.891 "state": "online", 00:32:01.891 "raid_level": "raid1", 00:32:01.891 "superblock": true, 00:32:01.891 "num_base_bdevs": 4, 00:32:01.891 "num_base_bdevs_discovered": 2, 00:32:01.891 "num_base_bdevs_operational": 2, 00:32:01.891 "base_bdevs_list": [ 00:32:01.891 { 00:32:01.891 "name": null, 00:32:01.891 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:01.891 "is_configured": false, 00:32:01.891 "data_offset": 0, 00:32:01.891 "data_size": 63488 00:32:01.891 }, 00:32:01.891 { 00:32:01.891 "name": null, 00:32:01.891 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:01.891 "is_configured": false, 00:32:01.891 "data_offset": 2048, 00:32:01.891 "data_size": 63488 00:32:01.891 }, 00:32:01.891 { 00:32:01.891 "name": "BaseBdev3", 00:32:01.891 "uuid": "59eda477-1d7f-5ac2-9bfb-d02bcbc4c870", 00:32:01.891 "is_configured": true, 00:32:01.891 "data_offset": 2048, 00:32:01.891 "data_size": 63488 00:32:01.891 }, 00:32:01.891 { 00:32:01.891 "name": "BaseBdev4", 00:32:01.891 "uuid": "d885740a-b18e-5b70-a7fd-311240bd474e", 00:32:01.891 "is_configured": true, 00:32:01.891 "data_offset": 2048, 00:32:01.891 "data_size": 63488 00:32:01.891 } 00:32:01.891 ] 00:32:01.891 }' 00:32:01.891 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:01.891 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:32:01.891 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:01.891 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:32:01.892 13:52:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 75797 00:32:01.892 13:52:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 75797 ']' 00:32:01.892 13:52:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 75797 00:32:01.892 13:52:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:32:01.892 13:52:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:32:01.892 13:52:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 75797 00:32:01.892 killing process with pid 75797 00:32:01.892 Received shutdown signal, test time was about 60.000000 seconds 00:32:01.892 00:32:01.892 Latency(us) 00:32:01.892 [2024-11-20T13:52:08.951Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:01.892 [2024-11-20T13:52:08.951Z] =================================================================================================================== 00:32:01.892 [2024-11-20T13:52:08.951Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:32:01.892 13:52:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:32:01.892 13:52:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:32:01.892 13:52:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 75797' 00:32:01.892 13:52:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 75797 00:32:01.892 [2024-11-20 13:52:08.761271] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:32:01.892 13:52:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 75797 00:32:01.892 [2024-11-20 13:52:08.761389] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:32:01.892 [2024-11-20 13:52:08.761456] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:32:01.892 [2024-11-20 13:52:08.761465] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:32:02.149 [2024-11-20 13:52:09.017711] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:32:02.714 13:52:09 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:32:02.714 ************************************ 00:32:02.714 END TEST raid_rebuild_test_sb 00:32:02.714 ************************************ 00:32:02.714 00:32:02.714 real 0m22.699s 00:32:02.714 user 0m26.289s 00:32:02.714 sys 0m3.252s 00:32:02.714 13:52:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:32:02.714 13:52:09 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:02.714 13:52:09 bdev_raid -- bdev/bdev_raid.sh@980 -- # run_test raid_rebuild_test_io raid_rebuild_test raid1 4 false true true 00:32:02.714 13:52:09 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:32:02.714 13:52:09 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:32:02.714 13:52:09 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:32:02.714 ************************************ 00:32:02.714 START TEST raid_rebuild_test_io 00:32:02.714 ************************************ 00:32:02.714 13:52:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 false true true 00:32:02.714 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:32:02.714 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:32:02.714 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:32:02.714 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:32:02.714 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:32:02.714 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:32:02.714 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@597 -- # raid_pid=76532 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 76532 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@835 -- # '[' -z 76532 ']' 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:02.715 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:32:02.715 13:52:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:02.715 [2024-11-20 13:52:09.760959] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:32:02.715 I/O size of 3145728 is greater than zero copy threshold (65536). 00:32:02.715 Zero copy mechanism will not be used. 00:32:02.715 [2024-11-20 13:52:09.761241] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid76532 ] 00:32:02.973 [2024-11-20 13:52:09.908010] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:02.973 [2024-11-20 13:52:10.009607] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:32:03.230 [2024-11-20 13:52:10.132776] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:32:03.230 [2024-11-20 13:52:10.132824] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:32:03.796 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:32:03.796 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # return 0 00:32:03.796 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:32:03.796 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:32:03.796 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:03.796 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:03.796 BaseBdev1_malloc 00:32:03.796 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:03.796 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:32:03.796 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:03.796 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:03.796 [2024-11-20 13:52:10.600777] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:32:03.796 [2024-11-20 13:52:10.600844] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:03.796 [2024-11-20 13:52:10.600866] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:32:03.796 [2024-11-20 13:52:10.600877] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:03.796 [2024-11-20 13:52:10.602832] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:03.796 [2024-11-20 13:52:10.602868] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:32:03.796 BaseBdev1 00:32:03.796 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:03.796 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:32:03.796 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:32:03.796 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:03.796 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:03.796 BaseBdev2_malloc 00:32:03.796 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:03.796 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:32:03.796 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:03.796 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:03.796 [2024-11-20 13:52:10.635061] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:32:03.796 [2024-11-20 13:52:10.635115] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:03.796 [2024-11-20 13:52:10.635133] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:32:03.797 [2024-11-20 13:52:10.635143] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:03.797 [2024-11-20 13:52:10.636957] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:03.797 [2024-11-20 13:52:10.636994] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:32:03.797 BaseBdev2 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:03.797 BaseBdev3_malloc 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:03.797 [2024-11-20 13:52:10.683110] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:32:03.797 [2024-11-20 13:52:10.683316] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:03.797 [2024-11-20 13:52:10.683342] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:32:03.797 [2024-11-20 13:52:10.683353] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:03.797 [2024-11-20 13:52:10.685251] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:03.797 [2024-11-20 13:52:10.685278] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:32:03.797 BaseBdev3 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:03.797 BaseBdev4_malloc 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:03.797 [2024-11-20 13:52:10.716656] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:32:03.797 [2024-11-20 13:52:10.716705] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:03.797 [2024-11-20 13:52:10.716720] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:32:03.797 [2024-11-20 13:52:10.716729] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:03.797 [2024-11-20 13:52:10.718544] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:03.797 [2024-11-20 13:52:10.718668] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:32:03.797 BaseBdev4 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:03.797 spare_malloc 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:03.797 spare_delay 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:03.797 [2024-11-20 13:52:10.757750] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:32:03.797 [2024-11-20 13:52:10.757888] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:03.797 [2024-11-20 13:52:10.757907] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:32:03.797 [2024-11-20 13:52:10.757917] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:03.797 [2024-11-20 13:52:10.759713] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:03.797 [2024-11-20 13:52:10.759739] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:32:03.797 spare 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:03.797 [2024-11-20 13:52:10.765790] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:32:03.797 [2024-11-20 13:52:10.767327] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:32:03.797 [2024-11-20 13:52:10.767374] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:32:03.797 [2024-11-20 13:52:10.767413] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:32:03.797 [2024-11-20 13:52:10.767476] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:32:03.797 [2024-11-20 13:52:10.767487] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:32:03.797 [2024-11-20 13:52:10.767701] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:32:03.797 [2024-11-20 13:52:10.767826] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:32:03.797 [2024-11-20 13:52:10.767835] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:32:03.797 [2024-11-20 13:52:10.767949] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:03.797 "name": "raid_bdev1", 00:32:03.797 "uuid": "ee385965-b3d5-4d38-9aa9-9551e15e7de5", 00:32:03.797 "strip_size_kb": 0, 00:32:03.797 "state": "online", 00:32:03.797 "raid_level": "raid1", 00:32:03.797 "superblock": false, 00:32:03.797 "num_base_bdevs": 4, 00:32:03.797 "num_base_bdevs_discovered": 4, 00:32:03.797 "num_base_bdevs_operational": 4, 00:32:03.797 "base_bdevs_list": [ 00:32:03.797 { 00:32:03.797 "name": "BaseBdev1", 00:32:03.797 "uuid": "26c2284d-2255-5414-b33a-928e8ce8bf50", 00:32:03.797 "is_configured": true, 00:32:03.797 "data_offset": 0, 00:32:03.797 "data_size": 65536 00:32:03.797 }, 00:32:03.797 { 00:32:03.797 "name": "BaseBdev2", 00:32:03.797 "uuid": "bd0aca18-3e7a-5d1f-a0b7-75c58411605e", 00:32:03.797 "is_configured": true, 00:32:03.797 "data_offset": 0, 00:32:03.797 "data_size": 65536 00:32:03.797 }, 00:32:03.797 { 00:32:03.797 "name": "BaseBdev3", 00:32:03.797 "uuid": "37365fce-0f0b-50e6-b5f3-fd515971bf79", 00:32:03.797 "is_configured": true, 00:32:03.797 "data_offset": 0, 00:32:03.797 "data_size": 65536 00:32:03.797 }, 00:32:03.797 { 00:32:03.797 "name": "BaseBdev4", 00:32:03.797 "uuid": "8e672b27-9423-5fcc-aec6-97438fdd4fdc", 00:32:03.797 "is_configured": true, 00:32:03.797 "data_offset": 0, 00:32:03.797 "data_size": 65536 00:32:03.797 } 00:32:03.797 ] 00:32:03.797 }' 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:03.797 13:52:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:04.055 13:52:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:32:04.055 13:52:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:32:04.055 13:52:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:04.055 13:52:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:04.055 [2024-11-20 13:52:11.106187] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:04.314 [2024-11-20 13:52:11.161836] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:04.314 "name": "raid_bdev1", 00:32:04.314 "uuid": "ee385965-b3d5-4d38-9aa9-9551e15e7de5", 00:32:04.314 "strip_size_kb": 0, 00:32:04.314 "state": "online", 00:32:04.314 "raid_level": "raid1", 00:32:04.314 "superblock": false, 00:32:04.314 "num_base_bdevs": 4, 00:32:04.314 "num_base_bdevs_discovered": 3, 00:32:04.314 "num_base_bdevs_operational": 3, 00:32:04.314 "base_bdevs_list": [ 00:32:04.314 { 00:32:04.314 "name": null, 00:32:04.314 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:04.314 "is_configured": false, 00:32:04.314 "data_offset": 0, 00:32:04.314 "data_size": 65536 00:32:04.314 }, 00:32:04.314 { 00:32:04.314 "name": "BaseBdev2", 00:32:04.314 "uuid": "bd0aca18-3e7a-5d1f-a0b7-75c58411605e", 00:32:04.314 "is_configured": true, 00:32:04.314 "data_offset": 0, 00:32:04.314 "data_size": 65536 00:32:04.314 }, 00:32:04.314 { 00:32:04.314 "name": "BaseBdev3", 00:32:04.314 "uuid": "37365fce-0f0b-50e6-b5f3-fd515971bf79", 00:32:04.314 "is_configured": true, 00:32:04.314 "data_offset": 0, 00:32:04.314 "data_size": 65536 00:32:04.314 }, 00:32:04.314 { 00:32:04.314 "name": "BaseBdev4", 00:32:04.314 "uuid": "8e672b27-9423-5fcc-aec6-97438fdd4fdc", 00:32:04.314 "is_configured": true, 00:32:04.314 "data_offset": 0, 00:32:04.314 "data_size": 65536 00:32:04.314 } 00:32:04.314 ] 00:32:04.314 }' 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:04.314 13:52:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:04.315 [2024-11-20 13:52:11.247023] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:32:04.315 I/O size of 3145728 is greater than zero copy threshold (65536). 00:32:04.315 Zero copy mechanism will not be used. 00:32:04.315 Running I/O for 60 seconds... 00:32:04.573 13:52:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:32:04.573 13:52:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:04.573 13:52:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:04.573 [2024-11-20 13:52:11.466339] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:32:04.573 13:52:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:04.573 13:52:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:32:04.573 [2024-11-20 13:52:11.511097] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000062f0 00:32:04.573 [2024-11-20 13:52:11.512968] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:32:04.573 [2024-11-20 13:52:11.618503] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:32:04.573 [2024-11-20 13:52:11.619155] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:32:04.831 [2024-11-20 13:52:11.835540] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:32:04.831 [2024-11-20 13:52:11.836364] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:32:05.397 [2024-11-20 13:52:12.184619] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:32:05.397 [2024-11-20 13:52:12.185355] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:32:05.397 176.00 IOPS, 528.00 MiB/s [2024-11-20T13:52:12.456Z] [2024-11-20 13:52:12.409919] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:32:05.656 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:05.657 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:05.657 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:05.657 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:05.657 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:05.657 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:05.657 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:05.657 13:52:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:05.657 13:52:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:05.657 13:52:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:05.657 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:05.657 "name": "raid_bdev1", 00:32:05.657 "uuid": "ee385965-b3d5-4d38-9aa9-9551e15e7de5", 00:32:05.657 "strip_size_kb": 0, 00:32:05.657 "state": "online", 00:32:05.657 "raid_level": "raid1", 00:32:05.657 "superblock": false, 00:32:05.657 "num_base_bdevs": 4, 00:32:05.657 "num_base_bdevs_discovered": 4, 00:32:05.657 "num_base_bdevs_operational": 4, 00:32:05.657 "process": { 00:32:05.657 "type": "rebuild", 00:32:05.657 "target": "spare", 00:32:05.657 "progress": { 00:32:05.657 "blocks": 10240, 00:32:05.657 "percent": 15 00:32:05.657 } 00:32:05.657 }, 00:32:05.657 "base_bdevs_list": [ 00:32:05.657 { 00:32:05.657 "name": "spare", 00:32:05.657 "uuid": "46778451-4ea5-5506-bbdb-b104287a43da", 00:32:05.657 "is_configured": true, 00:32:05.657 "data_offset": 0, 00:32:05.657 "data_size": 65536 00:32:05.657 }, 00:32:05.657 { 00:32:05.657 "name": "BaseBdev2", 00:32:05.657 "uuid": "bd0aca18-3e7a-5d1f-a0b7-75c58411605e", 00:32:05.657 "is_configured": true, 00:32:05.657 "data_offset": 0, 00:32:05.657 "data_size": 65536 00:32:05.657 }, 00:32:05.657 { 00:32:05.657 "name": "BaseBdev3", 00:32:05.657 "uuid": "37365fce-0f0b-50e6-b5f3-fd515971bf79", 00:32:05.657 "is_configured": true, 00:32:05.657 "data_offset": 0, 00:32:05.657 "data_size": 65536 00:32:05.657 }, 00:32:05.657 { 00:32:05.657 "name": "BaseBdev4", 00:32:05.657 "uuid": "8e672b27-9423-5fcc-aec6-97438fdd4fdc", 00:32:05.657 "is_configured": true, 00:32:05.657 "data_offset": 0, 00:32:05.657 "data_size": 65536 00:32:05.657 } 00:32:05.657 ] 00:32:05.657 }' 00:32:05.657 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:05.657 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:05.657 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:05.657 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:05.657 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:32:05.657 13:52:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:05.657 13:52:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:05.657 [2024-11-20 13:52:12.604397] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:32:05.657 [2024-11-20 13:52:12.647620] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:32:05.657 [2024-11-20 13:52:12.672373] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:32:05.657 [2024-11-20 13:52:12.681772] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:05.657 [2024-11-20 13:52:12.681815] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:32:05.657 [2024-11-20 13:52:12.681829] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:32:05.657 [2024-11-20 13:52:12.702444] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006220 00:32:05.932 13:52:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:05.932 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:32:05.932 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:05.932 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:05.932 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:05.932 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:05.932 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:05.932 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:05.932 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:05.932 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:05.932 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:05.932 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:05.932 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:05.932 13:52:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:05.932 13:52:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:05.932 13:52:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:05.932 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:05.932 "name": "raid_bdev1", 00:32:05.932 "uuid": "ee385965-b3d5-4d38-9aa9-9551e15e7de5", 00:32:05.932 "strip_size_kb": 0, 00:32:05.932 "state": "online", 00:32:05.932 "raid_level": "raid1", 00:32:05.932 "superblock": false, 00:32:05.932 "num_base_bdevs": 4, 00:32:05.932 "num_base_bdevs_discovered": 3, 00:32:05.932 "num_base_bdevs_operational": 3, 00:32:05.932 "base_bdevs_list": [ 00:32:05.932 { 00:32:05.932 "name": null, 00:32:05.932 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:05.932 "is_configured": false, 00:32:05.933 "data_offset": 0, 00:32:05.933 "data_size": 65536 00:32:05.933 }, 00:32:05.933 { 00:32:05.933 "name": "BaseBdev2", 00:32:05.933 "uuid": "bd0aca18-3e7a-5d1f-a0b7-75c58411605e", 00:32:05.933 "is_configured": true, 00:32:05.933 "data_offset": 0, 00:32:05.933 "data_size": 65536 00:32:05.933 }, 00:32:05.933 { 00:32:05.933 "name": "BaseBdev3", 00:32:05.933 "uuid": "37365fce-0f0b-50e6-b5f3-fd515971bf79", 00:32:05.933 "is_configured": true, 00:32:05.933 "data_offset": 0, 00:32:05.933 "data_size": 65536 00:32:05.933 }, 00:32:05.933 { 00:32:05.933 "name": "BaseBdev4", 00:32:05.933 "uuid": "8e672b27-9423-5fcc-aec6-97438fdd4fdc", 00:32:05.933 "is_configured": true, 00:32:05.933 "data_offset": 0, 00:32:05.933 "data_size": 65536 00:32:05.933 } 00:32:05.933 ] 00:32:05.933 }' 00:32:05.933 13:52:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:05.933 13:52:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:06.256 13:52:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:32:06.256 13:52:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:06.256 13:52:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:32:06.256 13:52:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:32:06.256 13:52:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:06.256 13:52:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:06.256 13:52:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:06.256 13:52:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:06.256 13:52:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:06.256 13:52:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:06.256 13:52:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:06.256 "name": "raid_bdev1", 00:32:06.256 "uuid": "ee385965-b3d5-4d38-9aa9-9551e15e7de5", 00:32:06.256 "strip_size_kb": 0, 00:32:06.256 "state": "online", 00:32:06.256 "raid_level": "raid1", 00:32:06.256 "superblock": false, 00:32:06.256 "num_base_bdevs": 4, 00:32:06.256 "num_base_bdevs_discovered": 3, 00:32:06.256 "num_base_bdevs_operational": 3, 00:32:06.256 "base_bdevs_list": [ 00:32:06.256 { 00:32:06.256 "name": null, 00:32:06.256 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:06.256 "is_configured": false, 00:32:06.256 "data_offset": 0, 00:32:06.256 "data_size": 65536 00:32:06.256 }, 00:32:06.256 { 00:32:06.256 "name": "BaseBdev2", 00:32:06.256 "uuid": "bd0aca18-3e7a-5d1f-a0b7-75c58411605e", 00:32:06.256 "is_configured": true, 00:32:06.256 "data_offset": 0, 00:32:06.256 "data_size": 65536 00:32:06.256 }, 00:32:06.256 { 00:32:06.256 "name": "BaseBdev3", 00:32:06.256 "uuid": "37365fce-0f0b-50e6-b5f3-fd515971bf79", 00:32:06.256 "is_configured": true, 00:32:06.256 "data_offset": 0, 00:32:06.256 "data_size": 65536 00:32:06.256 }, 00:32:06.256 { 00:32:06.256 "name": "BaseBdev4", 00:32:06.256 "uuid": "8e672b27-9423-5fcc-aec6-97438fdd4fdc", 00:32:06.256 "is_configured": true, 00:32:06.256 "data_offset": 0, 00:32:06.256 "data_size": 65536 00:32:06.256 } 00:32:06.256 ] 00:32:06.256 }' 00:32:06.256 13:52:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:06.256 13:52:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:32:06.256 13:52:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:06.256 13:52:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:32:06.256 13:52:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:32:06.256 13:52:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:06.256 13:52:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:06.256 [2024-11-20 13:52:13.155108] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:32:06.256 13:52:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:06.256 13:52:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:32:06.256 [2024-11-20 13:52:13.205207] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:32:06.256 [2024-11-20 13:52:13.207180] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:32:06.514 188.50 IOPS, 565.50 MiB/s [2024-11-20T13:52:13.573Z] [2024-11-20 13:52:13.327638] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:32:06.514 [2024-11-20 13:52:13.328193] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:32:06.514 [2024-11-20 13:52:13.533314] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:32:06.514 [2024-11-20 13:52:13.533627] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:32:07.080 [2024-11-20 13:52:13.873006] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:32:07.080 [2024-11-20 13:52:14.110656] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:32:07.340 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:07.340 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:07.340 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:07.340 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:07.341 "name": "raid_bdev1", 00:32:07.341 "uuid": "ee385965-b3d5-4d38-9aa9-9551e15e7de5", 00:32:07.341 "strip_size_kb": 0, 00:32:07.341 "state": "online", 00:32:07.341 "raid_level": "raid1", 00:32:07.341 "superblock": false, 00:32:07.341 "num_base_bdevs": 4, 00:32:07.341 "num_base_bdevs_discovered": 4, 00:32:07.341 "num_base_bdevs_operational": 4, 00:32:07.341 "process": { 00:32:07.341 "type": "rebuild", 00:32:07.341 "target": "spare", 00:32:07.341 "progress": { 00:32:07.341 "blocks": 10240, 00:32:07.341 "percent": 15 00:32:07.341 } 00:32:07.341 }, 00:32:07.341 "base_bdevs_list": [ 00:32:07.341 { 00:32:07.341 "name": "spare", 00:32:07.341 "uuid": "46778451-4ea5-5506-bbdb-b104287a43da", 00:32:07.341 "is_configured": true, 00:32:07.341 "data_offset": 0, 00:32:07.341 "data_size": 65536 00:32:07.341 }, 00:32:07.341 { 00:32:07.341 "name": "BaseBdev2", 00:32:07.341 "uuid": "bd0aca18-3e7a-5d1f-a0b7-75c58411605e", 00:32:07.341 "is_configured": true, 00:32:07.341 "data_offset": 0, 00:32:07.341 "data_size": 65536 00:32:07.341 }, 00:32:07.341 { 00:32:07.341 "name": "BaseBdev3", 00:32:07.341 "uuid": "37365fce-0f0b-50e6-b5f3-fd515971bf79", 00:32:07.341 "is_configured": true, 00:32:07.341 "data_offset": 0, 00:32:07.341 "data_size": 65536 00:32:07.341 }, 00:32:07.341 { 00:32:07.341 "name": "BaseBdev4", 00:32:07.341 "uuid": "8e672b27-9423-5fcc-aec6-97438fdd4fdc", 00:32:07.341 "is_configured": true, 00:32:07.341 "data_offset": 0, 00:32:07.341 "data_size": 65536 00:32:07.341 } 00:32:07.341 ] 00:32:07.341 }' 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:07.341 157.67 IOPS, 473.00 MiB/s [2024-11-20T13:52:14.400Z] 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:07.341 [2024-11-20 13:52:14.289914] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:32:07.341 [2024-11-20 13:52:14.334652] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000006220 00:32:07.341 [2024-11-20 13:52:14.334686] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d0000063c0 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:07.341 "name": "raid_bdev1", 00:32:07.341 "uuid": "ee385965-b3d5-4d38-9aa9-9551e15e7de5", 00:32:07.341 "strip_size_kb": 0, 00:32:07.341 "state": "online", 00:32:07.341 "raid_level": "raid1", 00:32:07.341 "superblock": false, 00:32:07.341 "num_base_bdevs": 4, 00:32:07.341 "num_base_bdevs_discovered": 3, 00:32:07.341 "num_base_bdevs_operational": 3, 00:32:07.341 "process": { 00:32:07.341 "type": "rebuild", 00:32:07.341 "target": "spare", 00:32:07.341 "progress": { 00:32:07.341 "blocks": 12288, 00:32:07.341 "percent": 18 00:32:07.341 } 00:32:07.341 }, 00:32:07.341 "base_bdevs_list": [ 00:32:07.341 { 00:32:07.341 "name": "spare", 00:32:07.341 "uuid": "46778451-4ea5-5506-bbdb-b104287a43da", 00:32:07.341 "is_configured": true, 00:32:07.341 "data_offset": 0, 00:32:07.341 "data_size": 65536 00:32:07.341 }, 00:32:07.341 { 00:32:07.341 "name": null, 00:32:07.341 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:07.341 "is_configured": false, 00:32:07.341 "data_offset": 0, 00:32:07.341 "data_size": 65536 00:32:07.341 }, 00:32:07.341 { 00:32:07.341 "name": "BaseBdev3", 00:32:07.341 "uuid": "37365fce-0f0b-50e6-b5f3-fd515971bf79", 00:32:07.341 "is_configured": true, 00:32:07.341 "data_offset": 0, 00:32:07.341 "data_size": 65536 00:32:07.341 }, 00:32:07.341 { 00:32:07.341 "name": "BaseBdev4", 00:32:07.341 "uuid": "8e672b27-9423-5fcc-aec6-97438fdd4fdc", 00:32:07.341 "is_configured": true, 00:32:07.341 "data_offset": 0, 00:32:07.341 "data_size": 65536 00:32:07.341 } 00:32:07.341 ] 00:32:07.341 }' 00:32:07.341 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:07.601 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:07.601 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:07.601 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:07.601 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # local timeout=381 00:32:07.601 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:32:07.601 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:07.601 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:07.601 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:07.602 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:07.602 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:07.602 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:07.602 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:07.602 13:52:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:07.602 13:52:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:07.602 13:52:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:07.602 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:07.602 "name": "raid_bdev1", 00:32:07.602 "uuid": "ee385965-b3d5-4d38-9aa9-9551e15e7de5", 00:32:07.602 "strip_size_kb": 0, 00:32:07.602 "state": "online", 00:32:07.602 "raid_level": "raid1", 00:32:07.602 "superblock": false, 00:32:07.602 "num_base_bdevs": 4, 00:32:07.602 "num_base_bdevs_discovered": 3, 00:32:07.602 "num_base_bdevs_operational": 3, 00:32:07.602 "process": { 00:32:07.602 "type": "rebuild", 00:32:07.602 "target": "spare", 00:32:07.602 "progress": { 00:32:07.602 "blocks": 12288, 00:32:07.602 "percent": 18 00:32:07.602 } 00:32:07.602 }, 00:32:07.602 "base_bdevs_list": [ 00:32:07.602 { 00:32:07.602 "name": "spare", 00:32:07.602 "uuid": "46778451-4ea5-5506-bbdb-b104287a43da", 00:32:07.602 "is_configured": true, 00:32:07.602 "data_offset": 0, 00:32:07.602 "data_size": 65536 00:32:07.602 }, 00:32:07.602 { 00:32:07.602 "name": null, 00:32:07.602 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:07.602 "is_configured": false, 00:32:07.602 "data_offset": 0, 00:32:07.602 "data_size": 65536 00:32:07.602 }, 00:32:07.602 { 00:32:07.602 "name": "BaseBdev3", 00:32:07.602 "uuid": "37365fce-0f0b-50e6-b5f3-fd515971bf79", 00:32:07.602 "is_configured": true, 00:32:07.602 "data_offset": 0, 00:32:07.602 "data_size": 65536 00:32:07.602 }, 00:32:07.602 { 00:32:07.602 "name": "BaseBdev4", 00:32:07.602 "uuid": "8e672b27-9423-5fcc-aec6-97438fdd4fdc", 00:32:07.602 "is_configured": true, 00:32:07.602 "data_offset": 0, 00:32:07.602 "data_size": 65536 00:32:07.602 } 00:32:07.602 ] 00:32:07.602 }' 00:32:07.602 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:07.602 [2024-11-20 13:52:14.464083] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:32:07.602 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:07.602 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:07.602 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:07.602 13:52:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:32:07.602 [2024-11-20 13:52:14.586021] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:32:07.602 [2024-11-20 13:52:14.586331] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:32:07.861 [2024-11-20 13:52:14.802347] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:32:07.861 [2024-11-20 13:52:14.803272] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:32:08.427 [2024-11-20 13:52:15.228936] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:32:08.427 137.00 IOPS, 411.00 MiB/s [2024-11-20T13:52:15.486Z] [2024-11-20 13:52:15.443155] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:32:08.685 13:52:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:32:08.685 13:52:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:08.685 13:52:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:08.685 13:52:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:08.685 13:52:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:08.685 13:52:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:08.685 13:52:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:08.685 13:52:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:08.685 13:52:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:08.685 13:52:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:08.685 13:52:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:08.685 13:52:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:08.685 "name": "raid_bdev1", 00:32:08.685 "uuid": "ee385965-b3d5-4d38-9aa9-9551e15e7de5", 00:32:08.685 "strip_size_kb": 0, 00:32:08.685 "state": "online", 00:32:08.685 "raid_level": "raid1", 00:32:08.685 "superblock": false, 00:32:08.685 "num_base_bdevs": 4, 00:32:08.685 "num_base_bdevs_discovered": 3, 00:32:08.685 "num_base_bdevs_operational": 3, 00:32:08.685 "process": { 00:32:08.685 "type": "rebuild", 00:32:08.685 "target": "spare", 00:32:08.685 "progress": { 00:32:08.685 "blocks": 28672, 00:32:08.685 "percent": 43 00:32:08.685 } 00:32:08.685 }, 00:32:08.685 "base_bdevs_list": [ 00:32:08.685 { 00:32:08.685 "name": "spare", 00:32:08.685 "uuid": "46778451-4ea5-5506-bbdb-b104287a43da", 00:32:08.685 "is_configured": true, 00:32:08.685 "data_offset": 0, 00:32:08.685 "data_size": 65536 00:32:08.685 }, 00:32:08.685 { 00:32:08.685 "name": null, 00:32:08.685 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:08.685 "is_configured": false, 00:32:08.685 "data_offset": 0, 00:32:08.685 "data_size": 65536 00:32:08.685 }, 00:32:08.685 { 00:32:08.685 "name": "BaseBdev3", 00:32:08.685 "uuid": "37365fce-0f0b-50e6-b5f3-fd515971bf79", 00:32:08.685 "is_configured": true, 00:32:08.685 "data_offset": 0, 00:32:08.685 "data_size": 65536 00:32:08.685 }, 00:32:08.685 { 00:32:08.685 "name": "BaseBdev4", 00:32:08.685 "uuid": "8e672b27-9423-5fcc-aec6-97438fdd4fdc", 00:32:08.685 "is_configured": true, 00:32:08.685 "data_offset": 0, 00:32:08.685 "data_size": 65536 00:32:08.686 } 00:32:08.686 ] 00:32:08.686 }' 00:32:08.686 13:52:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:08.686 13:52:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:08.686 13:52:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:08.686 13:52:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:08.686 13:52:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:32:08.686 [2024-11-20 13:52:15.652396] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:32:08.686 [2024-11-20 13:52:15.653319] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:32:09.007 [2024-11-20 13:52:15.860459] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:32:09.008 [2024-11-20 13:52:15.860961] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:32:09.321 121.00 IOPS, 363.00 MiB/s [2024-11-20T13:52:16.380Z] [2024-11-20 13:52:16.272254] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:32:09.606 13:52:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:32:09.606 13:52:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:09.606 13:52:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:09.607 13:52:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:09.607 13:52:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:09.607 13:52:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:09.607 13:52:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:09.607 13:52:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:09.607 13:52:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:09.607 13:52:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:09.607 13:52:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:09.865 13:52:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:09.865 "name": "raid_bdev1", 00:32:09.865 "uuid": "ee385965-b3d5-4d38-9aa9-9551e15e7de5", 00:32:09.865 "strip_size_kb": 0, 00:32:09.865 "state": "online", 00:32:09.865 "raid_level": "raid1", 00:32:09.865 "superblock": false, 00:32:09.865 "num_base_bdevs": 4, 00:32:09.865 "num_base_bdevs_discovered": 3, 00:32:09.865 "num_base_bdevs_operational": 3, 00:32:09.865 "process": { 00:32:09.865 "type": "rebuild", 00:32:09.865 "target": "spare", 00:32:09.865 "progress": { 00:32:09.865 "blocks": 45056, 00:32:09.865 "percent": 68 00:32:09.865 } 00:32:09.865 }, 00:32:09.865 "base_bdevs_list": [ 00:32:09.865 { 00:32:09.865 "name": "spare", 00:32:09.865 "uuid": "46778451-4ea5-5506-bbdb-b104287a43da", 00:32:09.865 "is_configured": true, 00:32:09.865 "data_offset": 0, 00:32:09.865 "data_size": 65536 00:32:09.865 }, 00:32:09.865 { 00:32:09.865 "name": null, 00:32:09.865 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:09.865 "is_configured": false, 00:32:09.865 "data_offset": 0, 00:32:09.865 "data_size": 65536 00:32:09.865 }, 00:32:09.865 { 00:32:09.865 "name": "BaseBdev3", 00:32:09.865 "uuid": "37365fce-0f0b-50e6-b5f3-fd515971bf79", 00:32:09.865 "is_configured": true, 00:32:09.865 "data_offset": 0, 00:32:09.865 "data_size": 65536 00:32:09.865 }, 00:32:09.865 { 00:32:09.865 "name": "BaseBdev4", 00:32:09.865 "uuid": "8e672b27-9423-5fcc-aec6-97438fdd4fdc", 00:32:09.865 "is_configured": true, 00:32:09.865 "data_offset": 0, 00:32:09.865 "data_size": 65536 00:32:09.865 } 00:32:09.865 ] 00:32:09.865 }' 00:32:09.865 13:52:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:09.865 13:52:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:09.865 13:52:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:09.865 [2024-11-20 13:52:16.712004] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:32:09.865 13:52:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:09.865 13:52:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:32:10.686 107.00 IOPS, 321.00 MiB/s [2024-11-20T13:52:17.745Z] [2024-11-20 13:52:17.683123] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:32:10.944 13:52:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:32:10.944 13:52:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:10.944 13:52:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:10.944 13:52:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:10.944 13:52:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:10.944 13:52:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:10.944 13:52:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:10.944 13:52:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:10.944 13:52:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:10.944 13:52:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:10.944 13:52:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:10.944 [2024-11-20 13:52:17.783218] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:32:10.944 [2024-11-20 13:52:17.786019] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:10.944 13:52:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:10.944 "name": "raid_bdev1", 00:32:10.944 "uuid": "ee385965-b3d5-4d38-9aa9-9551e15e7de5", 00:32:10.944 "strip_size_kb": 0, 00:32:10.944 "state": "online", 00:32:10.944 "raid_level": "raid1", 00:32:10.944 "superblock": false, 00:32:10.944 "num_base_bdevs": 4, 00:32:10.944 "num_base_bdevs_discovered": 3, 00:32:10.944 "num_base_bdevs_operational": 3, 00:32:10.944 "process": { 00:32:10.944 "type": "rebuild", 00:32:10.944 "target": "spare", 00:32:10.944 "progress": { 00:32:10.944 "blocks": 65536, 00:32:10.944 "percent": 100 00:32:10.944 } 00:32:10.944 }, 00:32:10.944 "base_bdevs_list": [ 00:32:10.944 { 00:32:10.944 "name": "spare", 00:32:10.944 "uuid": "46778451-4ea5-5506-bbdb-b104287a43da", 00:32:10.944 "is_configured": true, 00:32:10.944 "data_offset": 0, 00:32:10.944 "data_size": 65536 00:32:10.944 }, 00:32:10.944 { 00:32:10.944 "name": null, 00:32:10.944 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:10.944 "is_configured": false, 00:32:10.944 "data_offset": 0, 00:32:10.944 "data_size": 65536 00:32:10.944 }, 00:32:10.944 { 00:32:10.944 "name": "BaseBdev3", 00:32:10.944 "uuid": "37365fce-0f0b-50e6-b5f3-fd515971bf79", 00:32:10.944 "is_configured": true, 00:32:10.944 "data_offset": 0, 00:32:10.944 "data_size": 65536 00:32:10.944 }, 00:32:10.944 { 00:32:10.944 "name": "BaseBdev4", 00:32:10.944 "uuid": "8e672b27-9423-5fcc-aec6-97438fdd4fdc", 00:32:10.944 "is_configured": true, 00:32:10.944 "data_offset": 0, 00:32:10.944 "data_size": 65536 00:32:10.944 } 00:32:10.944 ] 00:32:10.944 }' 00:32:10.944 13:52:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:10.944 13:52:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:10.944 13:52:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:10.944 13:52:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:10.944 13:52:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:32:12.130 96.29 IOPS, 288.86 MiB/s [2024-11-20T13:52:19.189Z] 13:52:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:12.130 "name": "raid_bdev1", 00:32:12.130 "uuid": "ee385965-b3d5-4d38-9aa9-9551e15e7de5", 00:32:12.130 "strip_size_kb": 0, 00:32:12.130 "state": "online", 00:32:12.130 "raid_level": "raid1", 00:32:12.130 "superblock": false, 00:32:12.130 "num_base_bdevs": 4, 00:32:12.130 "num_base_bdevs_discovered": 3, 00:32:12.130 "num_base_bdevs_operational": 3, 00:32:12.130 "base_bdevs_list": [ 00:32:12.130 { 00:32:12.130 "name": "spare", 00:32:12.130 "uuid": "46778451-4ea5-5506-bbdb-b104287a43da", 00:32:12.130 "is_configured": true, 00:32:12.130 "data_offset": 0, 00:32:12.130 "data_size": 65536 00:32:12.130 }, 00:32:12.130 { 00:32:12.130 "name": null, 00:32:12.130 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:12.130 "is_configured": false, 00:32:12.130 "data_offset": 0, 00:32:12.130 "data_size": 65536 00:32:12.130 }, 00:32:12.130 { 00:32:12.130 "name": "BaseBdev3", 00:32:12.130 "uuid": "37365fce-0f0b-50e6-b5f3-fd515971bf79", 00:32:12.130 "is_configured": true, 00:32:12.130 "data_offset": 0, 00:32:12.130 "data_size": 65536 00:32:12.130 }, 00:32:12.130 { 00:32:12.130 "name": "BaseBdev4", 00:32:12.130 "uuid": "8e672b27-9423-5fcc-aec6-97438fdd4fdc", 00:32:12.130 "is_configured": true, 00:32:12.130 "data_offset": 0, 00:32:12.130 "data_size": 65536 00:32:12.130 } 00:32:12.130 ] 00:32:12.130 }' 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@709 -- # break 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:12.130 13:52:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:12.130 "name": "raid_bdev1", 00:32:12.130 "uuid": "ee385965-b3d5-4d38-9aa9-9551e15e7de5", 00:32:12.130 "strip_size_kb": 0, 00:32:12.130 "state": "online", 00:32:12.130 "raid_level": "raid1", 00:32:12.130 "superblock": false, 00:32:12.130 "num_base_bdevs": 4, 00:32:12.130 "num_base_bdevs_discovered": 3, 00:32:12.130 "num_base_bdevs_operational": 3, 00:32:12.130 "base_bdevs_list": [ 00:32:12.130 { 00:32:12.130 "name": "spare", 00:32:12.130 "uuid": "46778451-4ea5-5506-bbdb-b104287a43da", 00:32:12.130 "is_configured": true, 00:32:12.130 "data_offset": 0, 00:32:12.130 "data_size": 65536 00:32:12.130 }, 00:32:12.130 { 00:32:12.130 "name": null, 00:32:12.130 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:12.130 "is_configured": false, 00:32:12.130 "data_offset": 0, 00:32:12.130 "data_size": 65536 00:32:12.130 }, 00:32:12.130 { 00:32:12.130 "name": "BaseBdev3", 00:32:12.130 "uuid": "37365fce-0f0b-50e6-b5f3-fd515971bf79", 00:32:12.130 "is_configured": true, 00:32:12.131 "data_offset": 0, 00:32:12.131 "data_size": 65536 00:32:12.131 }, 00:32:12.131 { 00:32:12.131 "name": "BaseBdev4", 00:32:12.131 "uuid": "8e672b27-9423-5fcc-aec6-97438fdd4fdc", 00:32:12.131 "is_configured": true, 00:32:12.131 "data_offset": 0, 00:32:12.131 "data_size": 65536 00:32:12.131 } 00:32:12.131 ] 00:32:12.131 }' 00:32:12.131 13:52:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:12.131 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:32:12.131 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:12.131 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:32:12.131 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:32:12.131 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:12.131 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:12.131 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:12.131 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:12.131 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:12.131 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:12.131 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:12.131 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:12.131 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:12.131 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:12.131 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:12.131 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:12.131 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:12.131 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:12.131 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:12.131 "name": "raid_bdev1", 00:32:12.131 "uuid": "ee385965-b3d5-4d38-9aa9-9551e15e7de5", 00:32:12.131 "strip_size_kb": 0, 00:32:12.131 "state": "online", 00:32:12.131 "raid_level": "raid1", 00:32:12.131 "superblock": false, 00:32:12.131 "num_base_bdevs": 4, 00:32:12.131 "num_base_bdevs_discovered": 3, 00:32:12.131 "num_base_bdevs_operational": 3, 00:32:12.131 "base_bdevs_list": [ 00:32:12.131 { 00:32:12.131 "name": "spare", 00:32:12.131 "uuid": "46778451-4ea5-5506-bbdb-b104287a43da", 00:32:12.131 "is_configured": true, 00:32:12.131 "data_offset": 0, 00:32:12.131 "data_size": 65536 00:32:12.131 }, 00:32:12.131 { 00:32:12.131 "name": null, 00:32:12.131 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:12.131 "is_configured": false, 00:32:12.131 "data_offset": 0, 00:32:12.131 "data_size": 65536 00:32:12.131 }, 00:32:12.131 { 00:32:12.131 "name": "BaseBdev3", 00:32:12.131 "uuid": "37365fce-0f0b-50e6-b5f3-fd515971bf79", 00:32:12.131 "is_configured": true, 00:32:12.131 "data_offset": 0, 00:32:12.131 "data_size": 65536 00:32:12.131 }, 00:32:12.131 { 00:32:12.131 "name": "BaseBdev4", 00:32:12.131 "uuid": "8e672b27-9423-5fcc-aec6-97438fdd4fdc", 00:32:12.131 "is_configured": true, 00:32:12.131 "data_offset": 0, 00:32:12.131 "data_size": 65536 00:32:12.131 } 00:32:12.131 ] 00:32:12.131 }' 00:32:12.131 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:12.131 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:12.387 89.25 IOPS, 267.75 MiB/s [2024-11-20T13:52:19.447Z] 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:32:12.388 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:12.388 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:12.388 [2024-11-20 13:52:19.363873] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:32:12.388 [2024-11-20 13:52:19.363909] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:32:12.388 00:32:12.388 Latency(us) 00:32:12.388 [2024-11-20T13:52:19.447Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:12.388 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:32:12.388 raid_bdev1 : 8.17 88.24 264.72 0.00 0.00 15876.41 247.34 116956.55 00:32:12.388 [2024-11-20T13:52:19.447Z] =================================================================================================================== 00:32:12.388 [2024-11-20T13:52:19.447Z] Total : 88.24 264.72 0.00 0.00 15876.41 247.34 116956.55 00:32:12.388 { 00:32:12.388 "results": [ 00:32:12.388 { 00:32:12.388 "job": "raid_bdev1", 00:32:12.388 "core_mask": "0x1", 00:32:12.388 "workload": "randrw", 00:32:12.388 "percentage": 50, 00:32:12.388 "status": "finished", 00:32:12.388 "queue_depth": 2, 00:32:12.388 "io_size": 3145728, 00:32:12.388 "runtime": 8.17084, 00:32:12.388 "iops": 88.24062152728483, 00:32:12.388 "mibps": 264.7218645818545, 00:32:12.388 "io_failed": 0, 00:32:12.388 "io_timeout": 0, 00:32:12.388 "avg_latency_us": 15876.409328923504, 00:32:12.388 "min_latency_us": 247.3353846153846, 00:32:12.388 "max_latency_us": 116956.55384615385 00:32:12.388 } 00:32:12.388 ], 00:32:12.388 "core_count": 1 00:32:12.388 } 00:32:12.388 [2024-11-20 13:52:19.432558] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:32:12.388 [2024-11-20 13:52:19.432625] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:12.388 [2024-11-20 13:52:19.432720] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:32:12.388 [2024-11-20 13:52:19.432729] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:32:12.388 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:12.388 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:12.388 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:12.388 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:12.388 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # jq length 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:32:12.646 /dev/nbd0 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:32:12.646 1+0 records in 00:32:12.646 1+0 records out 00:32:12.646 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000246986 s, 16.6 MB/s 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z '' ']' 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@728 -- # continue 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev3 ']' 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev3 /dev/nbd1 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev3') 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:32:12.646 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev3 /dev/nbd1 00:32:12.954 /dev/nbd1 00:32:12.954 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:32:12.954 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:32:12.954 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:32:12.954 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:32:12.954 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:32:12.954 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:32:12.954 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:32:12.954 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:32:12.954 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:32:12.954 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:32:12.954 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:32:12.954 1+0 records in 00:32:12.954 1+0 records out 00:32:12.954 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000226385 s, 18.1 MB/s 00:32:12.954 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:32:12.954 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:32:12.954 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:32:12.954 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:32:12.954 13:52:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:32:12.954 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:32:12.954 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:32:12.954 13:52:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:32:13.213 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:32:13.213 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:32:13.213 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:32:13.213 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:32:13.213 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:32:13.213 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:32:13.213 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:32:13.213 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:32:13.213 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:32:13.213 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:32:13.213 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:32:13.213 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:32:13.213 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev4 ']' 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev4 /dev/nbd1 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev4') 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev4 /dev/nbd1 00:32:13.471 /dev/nbd1 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:32:13.471 1+0 records in 00:32:13.471 1+0 records out 00:32:13.471 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000320076 s, 12.8 MB/s 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:32:13.471 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:32:13.728 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:32:13.728 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:32:13.728 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:32:13.729 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:32:13.729 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:32:13.729 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:32:13.729 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:32:13.988 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:32:13.988 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:32:13.988 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:32:13.988 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:32:13.988 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:32:13.988 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:32:13.988 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:32:13.988 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:32:13.988 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:32:13.988 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:32:13.988 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:32:13.988 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:32:13.988 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:32:13.988 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:32:13.988 13:52:20 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:32:13.988 13:52:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:32:13.988 13:52:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:32:13.988 13:52:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:32:13.988 13:52:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:32:13.988 13:52:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:32:13.988 13:52:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:32:13.988 13:52:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:32:13.988 13:52:21 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:32:13.988 13:52:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:32:13.988 13:52:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@784 -- # killprocess 76532 00:32:13.988 13:52:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@954 -- # '[' -z 76532 ']' 00:32:13.988 13:52:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@958 -- # kill -0 76532 00:32:13.988 13:52:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # uname 00:32:13.988 13:52:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:32:13.988 13:52:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 76532 00:32:13.989 13:52:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:32:13.989 13:52:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:32:13.989 killing process with pid 76532 00:32:13.989 13:52:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 76532' 00:32:13.989 Received shutdown signal, test time was about 9.793245 seconds 00:32:13.989 00:32:13.989 Latency(us) 00:32:13.989 [2024-11-20T13:52:21.048Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:13.989 [2024-11-20T13:52:21.048Z] =================================================================================================================== 00:32:13.989 [2024-11-20T13:52:21.048Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:32:13.989 13:52:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@973 -- # kill 76532 00:32:13.989 13:52:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@978 -- # wait 76532 00:32:13.989 [2024-11-20 13:52:21.042207] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:32:14.247 [2024-11-20 13:52:21.260834] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@786 -- # return 0 00:32:15.181 00:32:15.181 real 0m12.222s 00:32:15.181 user 0m14.999s 00:32:15.181 sys 0m1.344s 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:32:15.181 ************************************ 00:32:15.181 END TEST raid_rebuild_test_io 00:32:15.181 ************************************ 00:32:15.181 13:52:21 bdev_raid -- bdev/bdev_raid.sh@981 -- # run_test raid_rebuild_test_sb_io raid_rebuild_test raid1 4 true true true 00:32:15.181 13:52:21 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:32:15.181 13:52:21 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:32:15.181 13:52:21 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:32:15.181 ************************************ 00:32:15.181 START TEST raid_rebuild_test_sb_io 00:32:15.181 ************************************ 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 true true true 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:32:15.181 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:32:15.182 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:32:15.182 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@597 -- # raid_pid=76938 00:32:15.182 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 76938 00:32:15.182 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@835 -- # '[' -z 76938 ']' 00:32:15.182 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:15.182 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:32:15.182 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:15.182 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:15.182 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:32:15.182 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:15.182 13:52:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:32:15.182 I/O size of 3145728 is greater than zero copy threshold (65536). 00:32:15.182 Zero copy mechanism will not be used. 00:32:15.182 [2024-11-20 13:52:22.062368] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:32:15.182 [2024-11-20 13:52:22.062572] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid76938 ] 00:32:15.182 [2024-11-20 13:52:22.236085] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:15.440 [2024-11-20 13:52:22.339694] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:32:15.440 [2024-11-20 13:52:22.460442] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:32:15.440 [2024-11-20 13:52:22.460500] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:32:16.005 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:32:16.005 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # return 0 00:32:16.005 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:32:16.005 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:32:16.005 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.005 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:16.005 BaseBdev1_malloc 00:32:16.005 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.005 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:32:16.005 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.005 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:16.005 [2024-11-20 13:52:22.910353] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:32:16.005 [2024-11-20 13:52:22.910409] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:16.005 [2024-11-20 13:52:22.910429] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:32:16.005 [2024-11-20 13:52:22.910439] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:16.005 [2024-11-20 13:52:22.912287] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:16.005 [2024-11-20 13:52:22.912317] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:32:16.005 BaseBdev1 00:32:16.005 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.005 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:32:16.005 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:32:16.005 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.005 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:16.005 BaseBdev2_malloc 00:32:16.005 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.005 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:32:16.005 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.005 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:16.006 [2024-11-20 13:52:22.944644] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:32:16.006 [2024-11-20 13:52:22.944698] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:16.006 [2024-11-20 13:52:22.944717] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:32:16.006 [2024-11-20 13:52:22.944727] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:16.006 [2024-11-20 13:52:22.946571] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:16.006 [2024-11-20 13:52:22.946597] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:32:16.006 BaseBdev2 00:32:16.006 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.006 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:32:16.006 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:32:16.006 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.006 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:16.006 BaseBdev3_malloc 00:32:16.006 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.006 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:32:16.006 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.006 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:16.006 [2024-11-20 13:52:22.991723] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:32:16.006 [2024-11-20 13:52:22.991764] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:16.006 [2024-11-20 13:52:22.991782] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:32:16.006 [2024-11-20 13:52:22.991792] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:16.006 [2024-11-20 13:52:22.993607] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:16.006 [2024-11-20 13:52:22.993637] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:32:16.006 BaseBdev3 00:32:16.006 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.006 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:32:16.006 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:32:16.006 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.006 13:52:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:16.006 BaseBdev4_malloc 00:32:16.006 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.006 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:32:16.006 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.006 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:16.006 [2024-11-20 13:52:23.025521] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:32:16.006 [2024-11-20 13:52:23.025566] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:16.006 [2024-11-20 13:52:23.025582] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:32:16.006 [2024-11-20 13:52:23.025593] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:16.006 [2024-11-20 13:52:23.027424] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:16.006 [2024-11-20 13:52:23.027452] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:32:16.006 BaseBdev4 00:32:16.006 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.006 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:32:16.006 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.006 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:16.006 spare_malloc 00:32:16.006 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.006 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:32:16.006 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.006 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:16.265 spare_delay 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:16.265 [2024-11-20 13:52:23.067058] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:32:16.265 [2024-11-20 13:52:23.067095] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:16.265 [2024-11-20 13:52:23.067110] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:32:16.265 [2024-11-20 13:52:23.067120] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:16.265 [2024-11-20 13:52:23.068967] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:16.265 [2024-11-20 13:52:23.069005] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:32:16.265 spare 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:16.265 [2024-11-20 13:52:23.075134] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:32:16.265 [2024-11-20 13:52:23.076974] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:32:16.265 [2024-11-20 13:52:23.077046] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:32:16.265 [2024-11-20 13:52:23.077089] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:32:16.265 [2024-11-20 13:52:23.077239] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:32:16.265 [2024-11-20 13:52:23.077254] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:32:16.265 [2024-11-20 13:52:23.077464] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:32:16.265 [2024-11-20 13:52:23.077605] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:32:16.265 [2024-11-20 13:52:23.077618] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:32:16.265 [2024-11-20 13:52:23.077736] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:16.265 "name": "raid_bdev1", 00:32:16.265 "uuid": "f0679e1a-4159-4f00-9c98-97c129b61d0e", 00:32:16.265 "strip_size_kb": 0, 00:32:16.265 "state": "online", 00:32:16.265 "raid_level": "raid1", 00:32:16.265 "superblock": true, 00:32:16.265 "num_base_bdevs": 4, 00:32:16.265 "num_base_bdevs_discovered": 4, 00:32:16.265 "num_base_bdevs_operational": 4, 00:32:16.265 "base_bdevs_list": [ 00:32:16.265 { 00:32:16.265 "name": "BaseBdev1", 00:32:16.265 "uuid": "b29b6d54-fd83-5f91-837d-982914de7a96", 00:32:16.265 "is_configured": true, 00:32:16.265 "data_offset": 2048, 00:32:16.265 "data_size": 63488 00:32:16.265 }, 00:32:16.265 { 00:32:16.265 "name": "BaseBdev2", 00:32:16.265 "uuid": "fb995025-a726-57e0-9dbd-6955118d5c1c", 00:32:16.265 "is_configured": true, 00:32:16.265 "data_offset": 2048, 00:32:16.265 "data_size": 63488 00:32:16.265 }, 00:32:16.265 { 00:32:16.265 "name": "BaseBdev3", 00:32:16.265 "uuid": "c5d06a40-515b-55af-b5ec-3ad678d14226", 00:32:16.265 "is_configured": true, 00:32:16.265 "data_offset": 2048, 00:32:16.265 "data_size": 63488 00:32:16.265 }, 00:32:16.265 { 00:32:16.265 "name": "BaseBdev4", 00:32:16.265 "uuid": "32609520-b04b-50c5-885a-8572452f2d36", 00:32:16.265 "is_configured": true, 00:32:16.265 "data_offset": 2048, 00:32:16.265 "data_size": 63488 00:32:16.265 } 00:32:16.265 ] 00:32:16.265 }' 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:16.265 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:16.524 [2024-11-20 13:52:23.407507] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:16.524 [2024-11-20 13:52:23.459179] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:16.524 "name": "raid_bdev1", 00:32:16.524 "uuid": "f0679e1a-4159-4f00-9c98-97c129b61d0e", 00:32:16.524 "strip_size_kb": 0, 00:32:16.524 "state": "online", 00:32:16.524 "raid_level": "raid1", 00:32:16.524 "superblock": true, 00:32:16.524 "num_base_bdevs": 4, 00:32:16.524 "num_base_bdevs_discovered": 3, 00:32:16.524 "num_base_bdevs_operational": 3, 00:32:16.524 "base_bdevs_list": [ 00:32:16.524 { 00:32:16.524 "name": null, 00:32:16.524 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:16.524 "is_configured": false, 00:32:16.524 "data_offset": 0, 00:32:16.524 "data_size": 63488 00:32:16.524 }, 00:32:16.524 { 00:32:16.524 "name": "BaseBdev2", 00:32:16.524 "uuid": "fb995025-a726-57e0-9dbd-6955118d5c1c", 00:32:16.524 "is_configured": true, 00:32:16.524 "data_offset": 2048, 00:32:16.524 "data_size": 63488 00:32:16.524 }, 00:32:16.524 { 00:32:16.524 "name": "BaseBdev3", 00:32:16.524 "uuid": "c5d06a40-515b-55af-b5ec-3ad678d14226", 00:32:16.524 "is_configured": true, 00:32:16.524 "data_offset": 2048, 00:32:16.524 "data_size": 63488 00:32:16.524 }, 00:32:16.524 { 00:32:16.524 "name": "BaseBdev4", 00:32:16.524 "uuid": "32609520-b04b-50c5-885a-8572452f2d36", 00:32:16.524 "is_configured": true, 00:32:16.524 "data_offset": 2048, 00:32:16.524 "data_size": 63488 00:32:16.524 } 00:32:16.524 ] 00:32:16.524 }' 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:16.524 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:16.524 [2024-11-20 13:52:23.552091] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:32:16.524 I/O size of 3145728 is greater than zero copy threshold (65536). 00:32:16.524 Zero copy mechanism will not be used. 00:32:16.525 Running I/O for 60 seconds... 00:32:16.865 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:32:16.865 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.865 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:16.865 [2024-11-20 13:52:23.813826] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:32:16.865 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.865 13:52:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:32:16.865 [2024-11-20 13:52:23.886162] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000062f0 00:32:16.865 [2024-11-20 13:52:23.887947] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:32:17.124 [2024-11-20 13:52:24.004561] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:32:17.124 [2024-11-20 13:52:24.005109] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:32:17.124 [2024-11-20 13:52:24.130582] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:32:17.124 [2024-11-20 13:52:24.131272] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:32:17.690 [2024-11-20 13:52:24.483995] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:32:17.690 [2024-11-20 13:52:24.484560] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:32:17.690 134.00 IOPS, 402.00 MiB/s [2024-11-20T13:52:24.749Z] [2024-11-20 13:52:24.722354] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:32:17.690 [2024-11-20 13:52:24.723015] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:32:17.948 13:52:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:17.948 13:52:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:17.948 13:52:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:17.948 13:52:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:17.948 13:52:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:17.948 13:52:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:17.948 13:52:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:17.948 13:52:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:17.948 13:52:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:17.948 13:52:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:17.948 13:52:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:17.948 "name": "raid_bdev1", 00:32:17.948 "uuid": "f0679e1a-4159-4f00-9c98-97c129b61d0e", 00:32:17.948 "strip_size_kb": 0, 00:32:17.948 "state": "online", 00:32:17.948 "raid_level": "raid1", 00:32:17.948 "superblock": true, 00:32:17.948 "num_base_bdevs": 4, 00:32:17.948 "num_base_bdevs_discovered": 4, 00:32:17.948 "num_base_bdevs_operational": 4, 00:32:17.948 "process": { 00:32:17.948 "type": "rebuild", 00:32:17.948 "target": "spare", 00:32:17.948 "progress": { 00:32:17.948 "blocks": 10240, 00:32:17.948 "percent": 16 00:32:17.948 } 00:32:17.948 }, 00:32:17.948 "base_bdevs_list": [ 00:32:17.948 { 00:32:17.948 "name": "spare", 00:32:17.948 "uuid": "6ffde06f-7485-5e71-b01d-1c2359a0a782", 00:32:17.948 "is_configured": true, 00:32:17.948 "data_offset": 2048, 00:32:17.948 "data_size": 63488 00:32:17.948 }, 00:32:17.948 { 00:32:17.948 "name": "BaseBdev2", 00:32:17.948 "uuid": "fb995025-a726-57e0-9dbd-6955118d5c1c", 00:32:17.948 "is_configured": true, 00:32:17.948 "data_offset": 2048, 00:32:17.948 "data_size": 63488 00:32:17.948 }, 00:32:17.948 { 00:32:17.948 "name": "BaseBdev3", 00:32:17.948 "uuid": "c5d06a40-515b-55af-b5ec-3ad678d14226", 00:32:17.948 "is_configured": true, 00:32:17.948 "data_offset": 2048, 00:32:17.948 "data_size": 63488 00:32:17.948 }, 00:32:17.948 { 00:32:17.948 "name": "BaseBdev4", 00:32:17.948 "uuid": "32609520-b04b-50c5-885a-8572452f2d36", 00:32:17.948 "is_configured": true, 00:32:17.948 "data_offset": 2048, 00:32:17.948 "data_size": 63488 00:32:17.948 } 00:32:17.948 ] 00:32:17.948 }' 00:32:17.948 13:52:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:17.948 13:52:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:17.948 13:52:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:17.948 13:52:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:17.948 13:52:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:32:17.948 13:52:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:17.948 13:52:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:17.948 [2024-11-20 13:52:24.954974] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:32:18.207 [2024-11-20 13:52:25.179066] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:32:18.208 [2024-11-20 13:52:25.199232] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:18.208 [2024-11-20 13:52:25.199300] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:32:18.208 [2024-11-20 13:52:25.199316] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:32:18.208 [2024-11-20 13:52:25.219931] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006220 00:32:18.208 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:18.208 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:32:18.208 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:18.208 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:18.208 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:18.208 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:18.208 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:18.208 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:18.208 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:18.208 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:18.208 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:18.208 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:18.208 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:18.208 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:18.208 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:18.208 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:18.466 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:18.466 "name": "raid_bdev1", 00:32:18.467 "uuid": "f0679e1a-4159-4f00-9c98-97c129b61d0e", 00:32:18.467 "strip_size_kb": 0, 00:32:18.467 "state": "online", 00:32:18.467 "raid_level": "raid1", 00:32:18.467 "superblock": true, 00:32:18.467 "num_base_bdevs": 4, 00:32:18.467 "num_base_bdevs_discovered": 3, 00:32:18.467 "num_base_bdevs_operational": 3, 00:32:18.467 "base_bdevs_list": [ 00:32:18.467 { 00:32:18.467 "name": null, 00:32:18.467 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:18.467 "is_configured": false, 00:32:18.467 "data_offset": 0, 00:32:18.467 "data_size": 63488 00:32:18.467 }, 00:32:18.467 { 00:32:18.467 "name": "BaseBdev2", 00:32:18.467 "uuid": "fb995025-a726-57e0-9dbd-6955118d5c1c", 00:32:18.467 "is_configured": true, 00:32:18.467 "data_offset": 2048, 00:32:18.467 "data_size": 63488 00:32:18.467 }, 00:32:18.467 { 00:32:18.467 "name": "BaseBdev3", 00:32:18.467 "uuid": "c5d06a40-515b-55af-b5ec-3ad678d14226", 00:32:18.467 "is_configured": true, 00:32:18.467 "data_offset": 2048, 00:32:18.467 "data_size": 63488 00:32:18.467 }, 00:32:18.467 { 00:32:18.467 "name": "BaseBdev4", 00:32:18.467 "uuid": "32609520-b04b-50c5-885a-8572452f2d36", 00:32:18.467 "is_configured": true, 00:32:18.467 "data_offset": 2048, 00:32:18.467 "data_size": 63488 00:32:18.467 } 00:32:18.467 ] 00:32:18.467 }' 00:32:18.467 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:18.467 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:18.725 120.50 IOPS, 361.50 MiB/s [2024-11-20T13:52:25.784Z] 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:32:18.725 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:18.725 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:32:18.725 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:32:18.725 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:18.725 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:18.725 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:18.725 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:18.725 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:18.725 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:18.725 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:18.725 "name": "raid_bdev1", 00:32:18.725 "uuid": "f0679e1a-4159-4f00-9c98-97c129b61d0e", 00:32:18.725 "strip_size_kb": 0, 00:32:18.725 "state": "online", 00:32:18.725 "raid_level": "raid1", 00:32:18.725 "superblock": true, 00:32:18.725 "num_base_bdevs": 4, 00:32:18.725 "num_base_bdevs_discovered": 3, 00:32:18.725 "num_base_bdevs_operational": 3, 00:32:18.725 "base_bdevs_list": [ 00:32:18.725 { 00:32:18.725 "name": null, 00:32:18.725 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:18.725 "is_configured": false, 00:32:18.725 "data_offset": 0, 00:32:18.725 "data_size": 63488 00:32:18.725 }, 00:32:18.725 { 00:32:18.725 "name": "BaseBdev2", 00:32:18.725 "uuid": "fb995025-a726-57e0-9dbd-6955118d5c1c", 00:32:18.725 "is_configured": true, 00:32:18.725 "data_offset": 2048, 00:32:18.725 "data_size": 63488 00:32:18.725 }, 00:32:18.725 { 00:32:18.725 "name": "BaseBdev3", 00:32:18.725 "uuid": "c5d06a40-515b-55af-b5ec-3ad678d14226", 00:32:18.725 "is_configured": true, 00:32:18.725 "data_offset": 2048, 00:32:18.725 "data_size": 63488 00:32:18.725 }, 00:32:18.725 { 00:32:18.725 "name": "BaseBdev4", 00:32:18.725 "uuid": "32609520-b04b-50c5-885a-8572452f2d36", 00:32:18.725 "is_configured": true, 00:32:18.725 "data_offset": 2048, 00:32:18.725 "data_size": 63488 00:32:18.725 } 00:32:18.725 ] 00:32:18.725 }' 00:32:18.725 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:18.725 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:32:18.725 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:18.725 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:32:18.725 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:32:18.725 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:18.725 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:18.725 [2024-11-20 13:52:25.687028] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:32:18.725 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:18.725 13:52:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:32:18.725 [2024-11-20 13:52:25.764556] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:32:18.725 [2024-11-20 13:52:25.766329] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:32:18.983 [2024-11-20 13:52:25.894346] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:32:18.983 [2024-11-20 13:52:26.032193] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:32:18.983 [2024-11-20 13:52:26.037531] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:32:19.552 [2024-11-20 13:52:26.374260] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:32:19.552 [2024-11-20 13:52:26.374806] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:32:19.552 [2024-11-20 13:52:26.506469] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:32:19.552 [2024-11-20 13:52:26.507156] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:32:19.811 127.00 IOPS, 381.00 MiB/s [2024-11-20T13:52:26.870Z] 13:52:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:19.811 13:52:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:19.811 13:52:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:19.811 13:52:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:19.811 13:52:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:19.811 13:52:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:19.811 13:52:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:19.811 13:52:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:19.811 13:52:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:19.811 13:52:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:19.811 13:52:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:19.811 "name": "raid_bdev1", 00:32:19.811 "uuid": "f0679e1a-4159-4f00-9c98-97c129b61d0e", 00:32:19.811 "strip_size_kb": 0, 00:32:19.811 "state": "online", 00:32:19.811 "raid_level": "raid1", 00:32:19.811 "superblock": true, 00:32:19.811 "num_base_bdevs": 4, 00:32:19.811 "num_base_bdevs_discovered": 4, 00:32:19.811 "num_base_bdevs_operational": 4, 00:32:19.811 "process": { 00:32:19.811 "type": "rebuild", 00:32:19.811 "target": "spare", 00:32:19.811 "progress": { 00:32:19.811 "blocks": 12288, 00:32:19.811 "percent": 19 00:32:19.811 } 00:32:19.811 }, 00:32:19.811 "base_bdevs_list": [ 00:32:19.811 { 00:32:19.811 "name": "spare", 00:32:19.811 "uuid": "6ffde06f-7485-5e71-b01d-1c2359a0a782", 00:32:19.811 "is_configured": true, 00:32:19.811 "data_offset": 2048, 00:32:19.811 "data_size": 63488 00:32:19.811 }, 00:32:19.811 { 00:32:19.811 "name": "BaseBdev2", 00:32:19.811 "uuid": "fb995025-a726-57e0-9dbd-6955118d5c1c", 00:32:19.811 "is_configured": true, 00:32:19.811 "data_offset": 2048, 00:32:19.811 "data_size": 63488 00:32:19.811 }, 00:32:19.811 { 00:32:19.811 "name": "BaseBdev3", 00:32:19.811 "uuid": "c5d06a40-515b-55af-b5ec-3ad678d14226", 00:32:19.811 "is_configured": true, 00:32:19.811 "data_offset": 2048, 00:32:19.811 "data_size": 63488 00:32:19.811 }, 00:32:19.811 { 00:32:19.811 "name": "BaseBdev4", 00:32:19.811 "uuid": "32609520-b04b-50c5-885a-8572452f2d36", 00:32:19.811 "is_configured": true, 00:32:19.811 "data_offset": 2048, 00:32:19.811 "data_size": 63488 00:32:19.811 } 00:32:19.811 ] 00:32:19.811 }' 00:32:19.811 13:52:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:19.811 13:52:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:19.811 13:52:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:19.811 13:52:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:19.811 13:52:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:32:19.811 13:52:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:32:19.811 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:32:19.811 13:52:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:32:19.811 13:52:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:32:19.811 13:52:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:32:19.811 13:52:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:32:19.811 13:52:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:19.811 13:52:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:19.811 [2024-11-20 13:52:26.835887] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:32:20.071 [2024-11-20 13:52:27.068373] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000006220 00:32:20.071 [2024-11-20 13:52:27.068439] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d0000063c0 00:32:20.071 [2024-11-20 13:52:27.068498] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:32:20.071 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:20.071 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:32:20.071 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:32:20.071 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:20.071 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:20.071 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:20.071 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:20.071 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:20.071 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:20.071 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:20.071 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:20.071 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:20.071 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:20.071 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:20.071 "name": "raid_bdev1", 00:32:20.071 "uuid": "f0679e1a-4159-4f00-9c98-97c129b61d0e", 00:32:20.071 "strip_size_kb": 0, 00:32:20.071 "state": "online", 00:32:20.071 "raid_level": "raid1", 00:32:20.071 "superblock": true, 00:32:20.071 "num_base_bdevs": 4, 00:32:20.071 "num_base_bdevs_discovered": 3, 00:32:20.071 "num_base_bdevs_operational": 3, 00:32:20.071 "process": { 00:32:20.071 "type": "rebuild", 00:32:20.071 "target": "spare", 00:32:20.071 "progress": { 00:32:20.071 "blocks": 14336, 00:32:20.071 "percent": 22 00:32:20.071 } 00:32:20.071 }, 00:32:20.071 "base_bdevs_list": [ 00:32:20.071 { 00:32:20.071 "name": "spare", 00:32:20.071 "uuid": "6ffde06f-7485-5e71-b01d-1c2359a0a782", 00:32:20.071 "is_configured": true, 00:32:20.071 "data_offset": 2048, 00:32:20.071 "data_size": 63488 00:32:20.071 }, 00:32:20.071 { 00:32:20.071 "name": null, 00:32:20.071 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:20.071 "is_configured": false, 00:32:20.071 "data_offset": 0, 00:32:20.071 "data_size": 63488 00:32:20.071 }, 00:32:20.071 { 00:32:20.071 "name": "BaseBdev3", 00:32:20.071 "uuid": "c5d06a40-515b-55af-b5ec-3ad678d14226", 00:32:20.071 "is_configured": true, 00:32:20.071 "data_offset": 2048, 00:32:20.071 "data_size": 63488 00:32:20.071 }, 00:32:20.071 { 00:32:20.071 "name": "BaseBdev4", 00:32:20.071 "uuid": "32609520-b04b-50c5-885a-8572452f2d36", 00:32:20.071 "is_configured": true, 00:32:20.071 "data_offset": 2048, 00:32:20.071 "data_size": 63488 00:32:20.071 } 00:32:20.071 ] 00:32:20.071 }' 00:32:20.071 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:20.330 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:20.330 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:20.330 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:20.330 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # local timeout=394 00:32:20.330 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:32:20.330 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:20.330 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:20.330 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:20.330 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:20.330 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:20.331 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:20.331 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:20.331 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:20.331 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:20.331 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:20.331 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:20.331 "name": "raid_bdev1", 00:32:20.331 "uuid": "f0679e1a-4159-4f00-9c98-97c129b61d0e", 00:32:20.331 "strip_size_kb": 0, 00:32:20.331 "state": "online", 00:32:20.331 "raid_level": "raid1", 00:32:20.331 "superblock": true, 00:32:20.331 "num_base_bdevs": 4, 00:32:20.331 "num_base_bdevs_discovered": 3, 00:32:20.331 "num_base_bdevs_operational": 3, 00:32:20.331 "process": { 00:32:20.331 "type": "rebuild", 00:32:20.331 "target": "spare", 00:32:20.331 "progress": { 00:32:20.331 "blocks": 14336, 00:32:20.331 "percent": 22 00:32:20.331 } 00:32:20.331 }, 00:32:20.331 "base_bdevs_list": [ 00:32:20.331 { 00:32:20.331 "name": "spare", 00:32:20.331 "uuid": "6ffde06f-7485-5e71-b01d-1c2359a0a782", 00:32:20.331 "is_configured": true, 00:32:20.331 "data_offset": 2048, 00:32:20.331 "data_size": 63488 00:32:20.331 }, 00:32:20.331 { 00:32:20.331 "name": null, 00:32:20.331 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:20.331 "is_configured": false, 00:32:20.331 "data_offset": 0, 00:32:20.331 "data_size": 63488 00:32:20.331 }, 00:32:20.331 { 00:32:20.331 "name": "BaseBdev3", 00:32:20.331 "uuid": "c5d06a40-515b-55af-b5ec-3ad678d14226", 00:32:20.331 "is_configured": true, 00:32:20.331 "data_offset": 2048, 00:32:20.331 "data_size": 63488 00:32:20.331 }, 00:32:20.331 { 00:32:20.331 "name": "BaseBdev4", 00:32:20.331 "uuid": "32609520-b04b-50c5-885a-8572452f2d36", 00:32:20.331 "is_configured": true, 00:32:20.331 "data_offset": 2048, 00:32:20.331 "data_size": 63488 00:32:20.331 } 00:32:20.331 ] 00:32:20.331 }' 00:32:20.331 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:20.331 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:20.331 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:20.331 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:20.331 13:52:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:32:20.331 [2024-11-20 13:52:27.310467] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:32:20.848 113.75 IOPS, 341.25 MiB/s [2024-11-20T13:52:27.907Z] [2024-11-20 13:52:27.661692] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:32:20.848 [2024-11-20 13:52:27.876932] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:32:21.414 13:52:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:32:21.414 13:52:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:21.414 13:52:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:21.414 13:52:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:21.414 13:52:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:21.414 13:52:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:21.414 13:52:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:21.414 13:52:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:21.414 13:52:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:21.414 13:52:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:21.414 13:52:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:21.414 13:52:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:21.414 "name": "raid_bdev1", 00:32:21.414 "uuid": "f0679e1a-4159-4f00-9c98-97c129b61d0e", 00:32:21.414 "strip_size_kb": 0, 00:32:21.414 "state": "online", 00:32:21.414 "raid_level": "raid1", 00:32:21.414 "superblock": true, 00:32:21.414 "num_base_bdevs": 4, 00:32:21.414 "num_base_bdevs_discovered": 3, 00:32:21.414 "num_base_bdevs_operational": 3, 00:32:21.414 "process": { 00:32:21.414 "type": "rebuild", 00:32:21.414 "target": "spare", 00:32:21.414 "progress": { 00:32:21.414 "blocks": 28672, 00:32:21.414 "percent": 45 00:32:21.414 } 00:32:21.414 }, 00:32:21.414 "base_bdevs_list": [ 00:32:21.414 { 00:32:21.414 "name": "spare", 00:32:21.414 "uuid": "6ffde06f-7485-5e71-b01d-1c2359a0a782", 00:32:21.414 "is_configured": true, 00:32:21.414 "data_offset": 2048, 00:32:21.414 "data_size": 63488 00:32:21.414 }, 00:32:21.414 { 00:32:21.414 "name": null, 00:32:21.414 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:21.414 "is_configured": false, 00:32:21.414 "data_offset": 0, 00:32:21.414 "data_size": 63488 00:32:21.414 }, 00:32:21.414 { 00:32:21.414 "name": "BaseBdev3", 00:32:21.414 "uuid": "c5d06a40-515b-55af-b5ec-3ad678d14226", 00:32:21.414 "is_configured": true, 00:32:21.414 "data_offset": 2048, 00:32:21.414 "data_size": 63488 00:32:21.414 }, 00:32:21.414 { 00:32:21.414 "name": "BaseBdev4", 00:32:21.414 "uuid": "32609520-b04b-50c5-885a-8572452f2d36", 00:32:21.414 "is_configured": true, 00:32:21.414 "data_offset": 2048, 00:32:21.414 "data_size": 63488 00:32:21.414 } 00:32:21.414 ] 00:32:21.414 }' 00:32:21.414 13:52:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:21.414 13:52:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:21.414 13:52:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:21.414 13:52:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:21.414 13:52:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:32:21.414 [2024-11-20 13:52:28.433514] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:32:22.605 102.80 IOPS, 308.40 MiB/s [2024-11-20T13:52:29.664Z] [2024-11-20 13:52:29.321025] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:32:22.605 13:52:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:32:22.605 13:52:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:22.605 13:52:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:22.605 13:52:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:22.605 13:52:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:22.605 13:52:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:22.605 13:52:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:22.606 13:52:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:22.606 13:52:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:22.606 13:52:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:22.606 13:52:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:22.606 13:52:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:22.606 "name": "raid_bdev1", 00:32:22.606 "uuid": "f0679e1a-4159-4f00-9c98-97c129b61d0e", 00:32:22.606 "strip_size_kb": 0, 00:32:22.606 "state": "online", 00:32:22.606 "raid_level": "raid1", 00:32:22.606 "superblock": true, 00:32:22.606 "num_base_bdevs": 4, 00:32:22.606 "num_base_bdevs_discovered": 3, 00:32:22.606 "num_base_bdevs_operational": 3, 00:32:22.606 "process": { 00:32:22.606 "type": "rebuild", 00:32:22.606 "target": "spare", 00:32:22.606 "progress": { 00:32:22.606 "blocks": 47104, 00:32:22.606 "percent": 74 00:32:22.606 } 00:32:22.606 }, 00:32:22.606 "base_bdevs_list": [ 00:32:22.606 { 00:32:22.606 "name": "spare", 00:32:22.606 "uuid": "6ffde06f-7485-5e71-b01d-1c2359a0a782", 00:32:22.606 "is_configured": true, 00:32:22.606 "data_offset": 2048, 00:32:22.606 "data_size": 63488 00:32:22.606 }, 00:32:22.606 { 00:32:22.606 "name": null, 00:32:22.606 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:22.606 "is_configured": false, 00:32:22.606 "data_offset": 0, 00:32:22.606 "data_size": 63488 00:32:22.606 }, 00:32:22.606 { 00:32:22.606 "name": "BaseBdev3", 00:32:22.606 "uuid": "c5d06a40-515b-55af-b5ec-3ad678d14226", 00:32:22.606 "is_configured": true, 00:32:22.606 "data_offset": 2048, 00:32:22.606 "data_size": 63488 00:32:22.606 }, 00:32:22.606 { 00:32:22.606 "name": "BaseBdev4", 00:32:22.606 "uuid": "32609520-b04b-50c5-885a-8572452f2d36", 00:32:22.606 "is_configured": true, 00:32:22.606 "data_offset": 2048, 00:32:22.606 "data_size": 63488 00:32:22.606 } 00:32:22.606 ] 00:32:22.606 }' 00:32:22.606 13:52:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:22.606 13:52:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:22.606 13:52:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:22.606 13:52:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:22.606 13:52:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:32:22.606 [2024-11-20 13:52:29.564514] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 51200 offset_begin: 49152 offset_end: 55296 00:32:23.542 93.33 IOPS, 280.00 MiB/s [2024-11-20T13:52:30.601Z] [2024-11-20 13:52:30.328517] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:32:23.542 [2024-11-20 13:52:30.433674] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:32:23.543 [2024-11-20 13:52:30.436279] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:23.543 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:32:23.543 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:23.543 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:23.543 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:23.543 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:23.543 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:23.543 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:23.543 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:23.543 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:23.543 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:23.543 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:23.543 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:23.543 "name": "raid_bdev1", 00:32:23.543 "uuid": "f0679e1a-4159-4f00-9c98-97c129b61d0e", 00:32:23.543 "strip_size_kb": 0, 00:32:23.543 "state": "online", 00:32:23.543 "raid_level": "raid1", 00:32:23.543 "superblock": true, 00:32:23.543 "num_base_bdevs": 4, 00:32:23.543 "num_base_bdevs_discovered": 3, 00:32:23.543 "num_base_bdevs_operational": 3, 00:32:23.543 "base_bdevs_list": [ 00:32:23.543 { 00:32:23.543 "name": "spare", 00:32:23.543 "uuid": "6ffde06f-7485-5e71-b01d-1c2359a0a782", 00:32:23.543 "is_configured": true, 00:32:23.543 "data_offset": 2048, 00:32:23.543 "data_size": 63488 00:32:23.543 }, 00:32:23.543 { 00:32:23.543 "name": null, 00:32:23.543 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:23.543 "is_configured": false, 00:32:23.543 "data_offset": 0, 00:32:23.543 "data_size": 63488 00:32:23.543 }, 00:32:23.543 { 00:32:23.543 "name": "BaseBdev3", 00:32:23.543 "uuid": "c5d06a40-515b-55af-b5ec-3ad678d14226", 00:32:23.543 "is_configured": true, 00:32:23.543 "data_offset": 2048, 00:32:23.543 "data_size": 63488 00:32:23.543 }, 00:32:23.543 { 00:32:23.543 "name": "BaseBdev4", 00:32:23.543 "uuid": "32609520-b04b-50c5-885a-8572452f2d36", 00:32:23.543 "is_configured": true, 00:32:23.543 "data_offset": 2048, 00:32:23.543 "data_size": 63488 00:32:23.543 } 00:32:23.543 ] 00:32:23.543 }' 00:32:23.543 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:23.543 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:32:23.543 86.00 IOPS, 258.00 MiB/s [2024-11-20T13:52:30.602Z] 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@709 -- # break 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:23.802 "name": "raid_bdev1", 00:32:23.802 "uuid": "f0679e1a-4159-4f00-9c98-97c129b61d0e", 00:32:23.802 "strip_size_kb": 0, 00:32:23.802 "state": "online", 00:32:23.802 "raid_level": "raid1", 00:32:23.802 "superblock": true, 00:32:23.802 "num_base_bdevs": 4, 00:32:23.802 "num_base_bdevs_discovered": 3, 00:32:23.802 "num_base_bdevs_operational": 3, 00:32:23.802 "base_bdevs_list": [ 00:32:23.802 { 00:32:23.802 "name": "spare", 00:32:23.802 "uuid": "6ffde06f-7485-5e71-b01d-1c2359a0a782", 00:32:23.802 "is_configured": true, 00:32:23.802 "data_offset": 2048, 00:32:23.802 "data_size": 63488 00:32:23.802 }, 00:32:23.802 { 00:32:23.802 "name": null, 00:32:23.802 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:23.802 "is_configured": false, 00:32:23.802 "data_offset": 0, 00:32:23.802 "data_size": 63488 00:32:23.802 }, 00:32:23.802 { 00:32:23.802 "name": "BaseBdev3", 00:32:23.802 "uuid": "c5d06a40-515b-55af-b5ec-3ad678d14226", 00:32:23.802 "is_configured": true, 00:32:23.802 "data_offset": 2048, 00:32:23.802 "data_size": 63488 00:32:23.802 }, 00:32:23.802 { 00:32:23.802 "name": "BaseBdev4", 00:32:23.802 "uuid": "32609520-b04b-50c5-885a-8572452f2d36", 00:32:23.802 "is_configured": true, 00:32:23.802 "data_offset": 2048, 00:32:23.802 "data_size": 63488 00:32:23.802 } 00:32:23.802 ] 00:32:23.802 }' 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:23.802 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:23.803 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:23.803 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:23.803 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:23.803 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:23.803 "name": "raid_bdev1", 00:32:23.803 "uuid": "f0679e1a-4159-4f00-9c98-97c129b61d0e", 00:32:23.803 "strip_size_kb": 0, 00:32:23.803 "state": "online", 00:32:23.803 "raid_level": "raid1", 00:32:23.803 "superblock": true, 00:32:23.803 "num_base_bdevs": 4, 00:32:23.803 "num_base_bdevs_discovered": 3, 00:32:23.803 "num_base_bdevs_operational": 3, 00:32:23.803 "base_bdevs_list": [ 00:32:23.803 { 00:32:23.803 "name": "spare", 00:32:23.803 "uuid": "6ffde06f-7485-5e71-b01d-1c2359a0a782", 00:32:23.803 "is_configured": true, 00:32:23.803 "data_offset": 2048, 00:32:23.803 "data_size": 63488 00:32:23.803 }, 00:32:23.803 { 00:32:23.803 "name": null, 00:32:23.803 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:23.803 "is_configured": false, 00:32:23.803 "data_offset": 0, 00:32:23.803 "data_size": 63488 00:32:23.803 }, 00:32:23.803 { 00:32:23.803 "name": "BaseBdev3", 00:32:23.803 "uuid": "c5d06a40-515b-55af-b5ec-3ad678d14226", 00:32:23.803 "is_configured": true, 00:32:23.803 "data_offset": 2048, 00:32:23.803 "data_size": 63488 00:32:23.803 }, 00:32:23.803 { 00:32:23.803 "name": "BaseBdev4", 00:32:23.803 "uuid": "32609520-b04b-50c5-885a-8572452f2d36", 00:32:23.803 "is_configured": true, 00:32:23.803 "data_offset": 2048, 00:32:23.803 "data_size": 63488 00:32:23.803 } 00:32:23.803 ] 00:32:23.803 }' 00:32:23.803 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:23.803 13:52:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:24.061 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:32:24.061 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:24.061 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:24.061 [2024-11-20 13:52:31.025438] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:32:24.061 [2024-11-20 13:52:31.025471] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:32:24.061 00:32:24.061 Latency(us) 00:32:24.061 [2024-11-20T13:52:31.120Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:24.061 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:32:24.061 raid_bdev1 : 7.54 82.15 246.44 0.00 0.00 16157.44 242.61 116956.55 00:32:24.061 [2024-11-20T13:52:31.121Z] =================================================================================================================== 00:32:24.062 [2024-11-20T13:52:31.121Z] Total : 82.15 246.44 0.00 0.00 16157.44 242.61 116956.55 00:32:24.062 [2024-11-20 13:52:31.101821] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:32:24.062 [2024-11-20 13:52:31.101895] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:24.062 [2024-11-20 13:52:31.102000] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:32:24.062 [2024-11-20 13:52:31.102012] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:32:24.062 { 00:32:24.062 "results": [ 00:32:24.062 { 00:32:24.062 "job": "raid_bdev1", 00:32:24.062 "core_mask": "0x1", 00:32:24.062 "workload": "randrw", 00:32:24.062 "percentage": 50, 00:32:24.062 "status": "finished", 00:32:24.062 "queue_depth": 2, 00:32:24.062 "io_size": 3145728, 00:32:24.062 "runtime": 7.535446, 00:32:24.062 "iops": 82.14510461623638, 00:32:24.062 "mibps": 246.43531384870914, 00:32:24.062 "io_failed": 0, 00:32:24.062 "io_timeout": 0, 00:32:24.062 "avg_latency_us": 16157.442385982353, 00:32:24.062 "min_latency_us": 242.60923076923078, 00:32:24.062 "max_latency_us": 116956.55384615385 00:32:24.062 } 00:32:24.062 ], 00:32:24.062 "core_count": 1 00:32:24.062 } 00:32:24.062 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:24.062 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:24.062 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # jq length 00:32:24.062 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:24.062 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:24.320 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:24.320 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:32:24.320 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:32:24.320 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:32:24.320 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:32:24.320 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:32:24.320 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:32:24.320 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:32:24.320 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:32:24.321 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:32:24.321 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:32:24.321 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:32:24.321 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:32:24.321 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:32:24.321 /dev/nbd0 00:32:24.321 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:32:24.321 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:32:24.321 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:32:24.321 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:32:24.321 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:32:24.321 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:32:24.321 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:32:24.321 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:32:24.321 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:32:24.321 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:32:24.321 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:32:24.321 1+0 records in 00:32:24.321 1+0 records out 00:32:24.321 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000272073 s, 15.1 MB/s 00:32:24.580 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:32:24.580 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:32:24.580 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:32:24.580 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:32:24.580 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:32:24.580 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:32:24.580 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:32:24.580 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:32:24.580 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z '' ']' 00:32:24.580 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@728 -- # continue 00:32:24.580 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:32:24.580 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev3 ']' 00:32:24.580 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev3 /dev/nbd1 00:32:24.580 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:32:24.580 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev3') 00:32:24.580 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:32:24.580 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:32:24.580 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:32:24.580 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:32:24.580 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:32:24.580 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:32:24.580 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev3 /dev/nbd1 00:32:24.580 /dev/nbd1 00:32:24.839 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:32:24.839 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:32:24.839 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:32:24.839 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:32:24.839 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:32:24.839 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:32:24.839 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:32:24.839 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:32:24.839 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:32:24.839 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:32:24.839 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:32:24.839 1+0 records in 00:32:24.839 1+0 records out 00:32:24.839 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000260807 s, 15.7 MB/s 00:32:24.839 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:32:24.839 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:32:24.839 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:32:24.839 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:32:24.839 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:32:24.839 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:32:24.839 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:32:24.839 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:32:24.839 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:32:24.839 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:32:24.839 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:32:24.839 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:32:24.839 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:32:24.839 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:32:24.839 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:32:25.098 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:32:25.098 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:32:25.098 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:32:25.098 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:32:25.098 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:32:25.098 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:32:25.098 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:32:25.098 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:32:25.098 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:32:25.098 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev4 ']' 00:32:25.098 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev4 /dev/nbd1 00:32:25.098 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:32:25.098 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev4') 00:32:25.098 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:32:25.098 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:32:25.098 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:32:25.098 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:32:25.098 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:32:25.098 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:32:25.098 13:52:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev4 /dev/nbd1 00:32:25.356 /dev/nbd1 00:32:25.356 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:32:25.356 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:32:25.356 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:32:25.356 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:32:25.356 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:32:25.356 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:32:25.356 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:32:25.356 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:32:25.356 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:32:25.356 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:32:25.356 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:32:25.356 1+0 records in 00:32:25.356 1+0 records out 00:32:25.356 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000357525 s, 11.5 MB/s 00:32:25.356 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:32:25.356 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:32:25.356 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:32:25.356 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:32:25.356 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:32:25.356 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:32:25.356 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:32:25.356 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:32:25.356 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:32:25.356 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:32:25.357 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:32:25.357 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:32:25.357 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:32:25.357 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:32:25.357 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:32:25.614 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:32:25.614 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:32:25.614 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:32:25.614 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:32:25.614 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:32:25.614 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:32:25.614 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:32:25.614 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:32:25.614 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:32:25.615 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:32:25.615 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:32:25.615 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:32:25.615 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:32:25.615 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:32:25.615 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:32:25.874 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:25.875 [2024-11-20 13:52:32.740375] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:32:25.875 [2024-11-20 13:52:32.740433] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:25.875 [2024-11-20 13:52:32.740453] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:32:25.875 [2024-11-20 13:52:32.740463] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:25.875 [2024-11-20 13:52:32.742487] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:25.875 [2024-11-20 13:52:32.742521] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:32:25.875 [2024-11-20 13:52:32.742603] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:32:25.875 [2024-11-20 13:52:32.742649] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:32:25.875 [2024-11-20 13:52:32.742768] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:32:25.875 [2024-11-20 13:52:32.742859] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:32:25.875 spare 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:25.875 [2024-11-20 13:52:32.842946] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:32:25.875 [2024-11-20 13:52:32.843015] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:32:25.875 [2024-11-20 13:52:32.843355] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000037160 00:32:25.875 [2024-11-20 13:52:32.843553] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:32:25.875 [2024-11-20 13:52:32.843567] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:32:25.875 [2024-11-20 13:52:32.843739] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:25.875 "name": "raid_bdev1", 00:32:25.875 "uuid": "f0679e1a-4159-4f00-9c98-97c129b61d0e", 00:32:25.875 "strip_size_kb": 0, 00:32:25.875 "state": "online", 00:32:25.875 "raid_level": "raid1", 00:32:25.875 "superblock": true, 00:32:25.875 "num_base_bdevs": 4, 00:32:25.875 "num_base_bdevs_discovered": 3, 00:32:25.875 "num_base_bdevs_operational": 3, 00:32:25.875 "base_bdevs_list": [ 00:32:25.875 { 00:32:25.875 "name": "spare", 00:32:25.875 "uuid": "6ffde06f-7485-5e71-b01d-1c2359a0a782", 00:32:25.875 "is_configured": true, 00:32:25.875 "data_offset": 2048, 00:32:25.875 "data_size": 63488 00:32:25.875 }, 00:32:25.875 { 00:32:25.875 "name": null, 00:32:25.875 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:25.875 "is_configured": false, 00:32:25.875 "data_offset": 2048, 00:32:25.875 "data_size": 63488 00:32:25.875 }, 00:32:25.875 { 00:32:25.875 "name": "BaseBdev3", 00:32:25.875 "uuid": "c5d06a40-515b-55af-b5ec-3ad678d14226", 00:32:25.875 "is_configured": true, 00:32:25.875 "data_offset": 2048, 00:32:25.875 "data_size": 63488 00:32:25.875 }, 00:32:25.875 { 00:32:25.875 "name": "BaseBdev4", 00:32:25.875 "uuid": "32609520-b04b-50c5-885a-8572452f2d36", 00:32:25.875 "is_configured": true, 00:32:25.875 "data_offset": 2048, 00:32:25.875 "data_size": 63488 00:32:25.875 } 00:32:25.875 ] 00:32:25.875 }' 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:25.875 13:52:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:26.134 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:32:26.134 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:26.134 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:32:26.134 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:32:26.134 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:26.134 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:26.134 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:26.134 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:26.134 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:26.134 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:26.134 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:26.134 "name": "raid_bdev1", 00:32:26.134 "uuid": "f0679e1a-4159-4f00-9c98-97c129b61d0e", 00:32:26.134 "strip_size_kb": 0, 00:32:26.134 "state": "online", 00:32:26.134 "raid_level": "raid1", 00:32:26.134 "superblock": true, 00:32:26.134 "num_base_bdevs": 4, 00:32:26.134 "num_base_bdevs_discovered": 3, 00:32:26.134 "num_base_bdevs_operational": 3, 00:32:26.134 "base_bdevs_list": [ 00:32:26.134 { 00:32:26.134 "name": "spare", 00:32:26.134 "uuid": "6ffde06f-7485-5e71-b01d-1c2359a0a782", 00:32:26.134 "is_configured": true, 00:32:26.135 "data_offset": 2048, 00:32:26.135 "data_size": 63488 00:32:26.135 }, 00:32:26.135 { 00:32:26.135 "name": null, 00:32:26.135 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:26.135 "is_configured": false, 00:32:26.135 "data_offset": 2048, 00:32:26.135 "data_size": 63488 00:32:26.135 }, 00:32:26.135 { 00:32:26.135 "name": "BaseBdev3", 00:32:26.135 "uuid": "c5d06a40-515b-55af-b5ec-3ad678d14226", 00:32:26.135 "is_configured": true, 00:32:26.135 "data_offset": 2048, 00:32:26.135 "data_size": 63488 00:32:26.135 }, 00:32:26.135 { 00:32:26.135 "name": "BaseBdev4", 00:32:26.135 "uuid": "32609520-b04b-50c5-885a-8572452f2d36", 00:32:26.135 "is_configured": true, 00:32:26.135 "data_offset": 2048, 00:32:26.135 "data_size": 63488 00:32:26.135 } 00:32:26.135 ] 00:32:26.135 }' 00:32:26.135 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:26.394 [2024-11-20 13:52:33.272612] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:26.394 "name": "raid_bdev1", 00:32:26.394 "uuid": "f0679e1a-4159-4f00-9c98-97c129b61d0e", 00:32:26.394 "strip_size_kb": 0, 00:32:26.394 "state": "online", 00:32:26.394 "raid_level": "raid1", 00:32:26.394 "superblock": true, 00:32:26.394 "num_base_bdevs": 4, 00:32:26.394 "num_base_bdevs_discovered": 2, 00:32:26.394 "num_base_bdevs_operational": 2, 00:32:26.394 "base_bdevs_list": [ 00:32:26.394 { 00:32:26.394 "name": null, 00:32:26.394 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:26.394 "is_configured": false, 00:32:26.394 "data_offset": 0, 00:32:26.394 "data_size": 63488 00:32:26.394 }, 00:32:26.394 { 00:32:26.394 "name": null, 00:32:26.394 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:26.394 "is_configured": false, 00:32:26.394 "data_offset": 2048, 00:32:26.394 "data_size": 63488 00:32:26.394 }, 00:32:26.394 { 00:32:26.394 "name": "BaseBdev3", 00:32:26.394 "uuid": "c5d06a40-515b-55af-b5ec-3ad678d14226", 00:32:26.394 "is_configured": true, 00:32:26.394 "data_offset": 2048, 00:32:26.394 "data_size": 63488 00:32:26.394 }, 00:32:26.394 { 00:32:26.394 "name": "BaseBdev4", 00:32:26.394 "uuid": "32609520-b04b-50c5-885a-8572452f2d36", 00:32:26.394 "is_configured": true, 00:32:26.394 "data_offset": 2048, 00:32:26.394 "data_size": 63488 00:32:26.394 } 00:32:26.394 ] 00:32:26.394 }' 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:26.394 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:26.654 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:32:26.654 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:26.654 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:26.654 [2024-11-20 13:52:33.608734] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:32:26.654 [2024-11-20 13:52:33.608928] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:32:26.654 [2024-11-20 13:52:33.608949] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:32:26.654 [2024-11-20 13:52:33.608997] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:32:26.654 [2024-11-20 13:52:33.617464] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000037230 00:32:26.654 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:26.654 13:52:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@757 -- # sleep 1 00:32:26.654 [2024-11-20 13:52:33.619227] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:32:27.590 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:27.590 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:27.590 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:27.590 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:27.590 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:27.590 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:27.590 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:27.590 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:27.590 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:27.590 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:27.857 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:27.857 "name": "raid_bdev1", 00:32:27.857 "uuid": "f0679e1a-4159-4f00-9c98-97c129b61d0e", 00:32:27.857 "strip_size_kb": 0, 00:32:27.857 "state": "online", 00:32:27.857 "raid_level": "raid1", 00:32:27.857 "superblock": true, 00:32:27.857 "num_base_bdevs": 4, 00:32:27.857 "num_base_bdevs_discovered": 3, 00:32:27.857 "num_base_bdevs_operational": 3, 00:32:27.857 "process": { 00:32:27.857 "type": "rebuild", 00:32:27.857 "target": "spare", 00:32:27.857 "progress": { 00:32:27.857 "blocks": 20480, 00:32:27.857 "percent": 32 00:32:27.857 } 00:32:27.857 }, 00:32:27.857 "base_bdevs_list": [ 00:32:27.857 { 00:32:27.857 "name": "spare", 00:32:27.857 "uuid": "6ffde06f-7485-5e71-b01d-1c2359a0a782", 00:32:27.857 "is_configured": true, 00:32:27.857 "data_offset": 2048, 00:32:27.857 "data_size": 63488 00:32:27.857 }, 00:32:27.857 { 00:32:27.857 "name": null, 00:32:27.857 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:27.857 "is_configured": false, 00:32:27.857 "data_offset": 2048, 00:32:27.857 "data_size": 63488 00:32:27.857 }, 00:32:27.857 { 00:32:27.857 "name": "BaseBdev3", 00:32:27.857 "uuid": "c5d06a40-515b-55af-b5ec-3ad678d14226", 00:32:27.857 "is_configured": true, 00:32:27.857 "data_offset": 2048, 00:32:27.857 "data_size": 63488 00:32:27.857 }, 00:32:27.857 { 00:32:27.857 "name": "BaseBdev4", 00:32:27.857 "uuid": "32609520-b04b-50c5-885a-8572452f2d36", 00:32:27.857 "is_configured": true, 00:32:27.857 "data_offset": 2048, 00:32:27.857 "data_size": 63488 00:32:27.857 } 00:32:27.857 ] 00:32:27.857 }' 00:32:27.857 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:27.857 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:27.857 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:27.857 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:27.857 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:32:27.857 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:27.857 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:27.857 [2024-11-20 13:52:34.729438] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:32:27.857 [2024-11-20 13:52:34.826464] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:32:27.857 [2024-11-20 13:52:34.826545] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:27.857 [2024-11-20 13:52:34.826559] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:32:27.857 [2024-11-20 13:52:34.826568] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:32:27.857 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:27.857 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:32:27.857 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:27.857 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:27.857 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:27.857 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:27.857 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:27.857 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:27.857 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:27.857 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:27.857 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:27.857 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:27.857 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:27.857 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:27.857 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:27.857 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:27.857 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:27.857 "name": "raid_bdev1", 00:32:27.857 "uuid": "f0679e1a-4159-4f00-9c98-97c129b61d0e", 00:32:27.857 "strip_size_kb": 0, 00:32:27.857 "state": "online", 00:32:27.857 "raid_level": "raid1", 00:32:27.857 "superblock": true, 00:32:27.857 "num_base_bdevs": 4, 00:32:27.857 "num_base_bdevs_discovered": 2, 00:32:27.857 "num_base_bdevs_operational": 2, 00:32:27.857 "base_bdevs_list": [ 00:32:27.857 { 00:32:27.857 "name": null, 00:32:27.857 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:27.857 "is_configured": false, 00:32:27.857 "data_offset": 0, 00:32:27.857 "data_size": 63488 00:32:27.857 }, 00:32:27.857 { 00:32:27.857 "name": null, 00:32:27.857 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:27.857 "is_configured": false, 00:32:27.857 "data_offset": 2048, 00:32:27.857 "data_size": 63488 00:32:27.857 }, 00:32:27.857 { 00:32:27.857 "name": "BaseBdev3", 00:32:27.857 "uuid": "c5d06a40-515b-55af-b5ec-3ad678d14226", 00:32:27.857 "is_configured": true, 00:32:27.857 "data_offset": 2048, 00:32:27.857 "data_size": 63488 00:32:27.857 }, 00:32:27.858 { 00:32:27.858 "name": "BaseBdev4", 00:32:27.858 "uuid": "32609520-b04b-50c5-885a-8572452f2d36", 00:32:27.858 "is_configured": true, 00:32:27.858 "data_offset": 2048, 00:32:27.858 "data_size": 63488 00:32:27.858 } 00:32:27.858 ] 00:32:27.858 }' 00:32:27.858 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:27.858 13:52:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:28.425 13:52:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:32:28.425 13:52:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:28.425 13:52:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:28.425 [2024-11-20 13:52:35.184746] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:32:28.425 [2024-11-20 13:52:35.184812] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:28.425 [2024-11-20 13:52:35.184839] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c680 00:32:28.425 [2024-11-20 13:52:35.184850] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:28.425 [2024-11-20 13:52:35.185314] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:28.425 [2024-11-20 13:52:35.185336] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:32:28.425 [2024-11-20 13:52:35.185424] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:32:28.425 [2024-11-20 13:52:35.185437] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:32:28.425 [2024-11-20 13:52:35.185446] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:32:28.425 [2024-11-20 13:52:35.185466] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:32:28.425 [2024-11-20 13:52:35.193946] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000037300 00:32:28.425 spare 00:32:28.425 13:52:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:28.425 13:52:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@764 -- # sleep 1 00:32:28.425 [2024-11-20 13:52:35.195670] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:32:29.360 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:29.360 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:29.360 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:29.360 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:29.360 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:29.360 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:29.360 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:29.360 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:29.360 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:29.360 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:29.360 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:29.360 "name": "raid_bdev1", 00:32:29.360 "uuid": "f0679e1a-4159-4f00-9c98-97c129b61d0e", 00:32:29.360 "strip_size_kb": 0, 00:32:29.360 "state": "online", 00:32:29.360 "raid_level": "raid1", 00:32:29.360 "superblock": true, 00:32:29.360 "num_base_bdevs": 4, 00:32:29.360 "num_base_bdevs_discovered": 3, 00:32:29.360 "num_base_bdevs_operational": 3, 00:32:29.360 "process": { 00:32:29.360 "type": "rebuild", 00:32:29.360 "target": "spare", 00:32:29.360 "progress": { 00:32:29.360 "blocks": 20480, 00:32:29.360 "percent": 32 00:32:29.360 } 00:32:29.360 }, 00:32:29.360 "base_bdevs_list": [ 00:32:29.360 { 00:32:29.360 "name": "spare", 00:32:29.360 "uuid": "6ffde06f-7485-5e71-b01d-1c2359a0a782", 00:32:29.360 "is_configured": true, 00:32:29.360 "data_offset": 2048, 00:32:29.360 "data_size": 63488 00:32:29.360 }, 00:32:29.360 { 00:32:29.360 "name": null, 00:32:29.360 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:29.360 "is_configured": false, 00:32:29.360 "data_offset": 2048, 00:32:29.360 "data_size": 63488 00:32:29.360 }, 00:32:29.360 { 00:32:29.360 "name": "BaseBdev3", 00:32:29.360 "uuid": "c5d06a40-515b-55af-b5ec-3ad678d14226", 00:32:29.360 "is_configured": true, 00:32:29.360 "data_offset": 2048, 00:32:29.360 "data_size": 63488 00:32:29.360 }, 00:32:29.360 { 00:32:29.360 "name": "BaseBdev4", 00:32:29.360 "uuid": "32609520-b04b-50c5-885a-8572452f2d36", 00:32:29.360 "is_configured": true, 00:32:29.360 "data_offset": 2048, 00:32:29.360 "data_size": 63488 00:32:29.360 } 00:32:29.360 ] 00:32:29.360 }' 00:32:29.360 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:29.360 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:29.360 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:29.360 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:29.360 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:32:29.360 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:29.360 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:29.360 [2024-11-20 13:52:36.301917] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:32:29.360 [2024-11-20 13:52:36.302338] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:32:29.360 [2024-11-20 13:52:36.302385] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:29.360 [2024-11-20 13:52:36.302400] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:32:29.361 [2024-11-20 13:52:36.302407] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:32:29.361 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:29.361 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:32:29.361 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:29.361 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:29.361 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:29.361 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:29.361 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:29.361 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:29.361 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:29.361 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:29.361 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:29.361 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:29.361 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:29.361 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:29.361 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:29.361 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:29.361 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:29.361 "name": "raid_bdev1", 00:32:29.361 "uuid": "f0679e1a-4159-4f00-9c98-97c129b61d0e", 00:32:29.361 "strip_size_kb": 0, 00:32:29.361 "state": "online", 00:32:29.361 "raid_level": "raid1", 00:32:29.361 "superblock": true, 00:32:29.361 "num_base_bdevs": 4, 00:32:29.361 "num_base_bdevs_discovered": 2, 00:32:29.361 "num_base_bdevs_operational": 2, 00:32:29.361 "base_bdevs_list": [ 00:32:29.361 { 00:32:29.361 "name": null, 00:32:29.361 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:29.361 "is_configured": false, 00:32:29.361 "data_offset": 0, 00:32:29.361 "data_size": 63488 00:32:29.361 }, 00:32:29.361 { 00:32:29.361 "name": null, 00:32:29.361 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:29.361 "is_configured": false, 00:32:29.361 "data_offset": 2048, 00:32:29.361 "data_size": 63488 00:32:29.361 }, 00:32:29.361 { 00:32:29.361 "name": "BaseBdev3", 00:32:29.361 "uuid": "c5d06a40-515b-55af-b5ec-3ad678d14226", 00:32:29.361 "is_configured": true, 00:32:29.361 "data_offset": 2048, 00:32:29.361 "data_size": 63488 00:32:29.361 }, 00:32:29.361 { 00:32:29.361 "name": "BaseBdev4", 00:32:29.361 "uuid": "32609520-b04b-50c5-885a-8572452f2d36", 00:32:29.361 "is_configured": true, 00:32:29.361 "data_offset": 2048, 00:32:29.361 "data_size": 63488 00:32:29.361 } 00:32:29.361 ] 00:32:29.361 }' 00:32:29.361 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:29.361 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:29.619 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:32:29.619 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:29.619 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:32:29.619 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:32:29.619 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:29.619 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:29.619 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:29.619 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:29.619 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:29.619 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:29.619 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:29.619 "name": "raid_bdev1", 00:32:29.619 "uuid": "f0679e1a-4159-4f00-9c98-97c129b61d0e", 00:32:29.619 "strip_size_kb": 0, 00:32:29.619 "state": "online", 00:32:29.619 "raid_level": "raid1", 00:32:29.619 "superblock": true, 00:32:29.619 "num_base_bdevs": 4, 00:32:29.619 "num_base_bdevs_discovered": 2, 00:32:29.619 "num_base_bdevs_operational": 2, 00:32:29.619 "base_bdevs_list": [ 00:32:29.619 { 00:32:29.619 "name": null, 00:32:29.619 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:29.619 "is_configured": false, 00:32:29.619 "data_offset": 0, 00:32:29.619 "data_size": 63488 00:32:29.619 }, 00:32:29.619 { 00:32:29.619 "name": null, 00:32:29.619 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:29.619 "is_configured": false, 00:32:29.619 "data_offset": 2048, 00:32:29.619 "data_size": 63488 00:32:29.619 }, 00:32:29.619 { 00:32:29.619 "name": "BaseBdev3", 00:32:29.619 "uuid": "c5d06a40-515b-55af-b5ec-3ad678d14226", 00:32:29.619 "is_configured": true, 00:32:29.619 "data_offset": 2048, 00:32:29.619 "data_size": 63488 00:32:29.619 }, 00:32:29.619 { 00:32:29.619 "name": "BaseBdev4", 00:32:29.619 "uuid": "32609520-b04b-50c5-885a-8572452f2d36", 00:32:29.619 "is_configured": true, 00:32:29.619 "data_offset": 2048, 00:32:29.619 "data_size": 63488 00:32:29.619 } 00:32:29.619 ] 00:32:29.619 }' 00:32:29.619 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:29.878 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:32:29.878 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:29.878 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:32:29.878 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:32:29.878 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:29.878 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:29.878 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:29.878 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:32:29.878 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:29.878 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:29.878 [2024-11-20 13:52:36.753010] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:32:29.878 [2024-11-20 13:52:36.753068] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:29.878 [2024-11-20 13:52:36.753088] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000cc80 00:32:29.878 [2024-11-20 13:52:36.753096] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:29.878 [2024-11-20 13:52:36.753532] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:29.878 [2024-11-20 13:52:36.753549] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:32:29.878 [2024-11-20 13:52:36.753622] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:32:29.878 [2024-11-20 13:52:36.753639] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:32:29.878 [2024-11-20 13:52:36.753648] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:32:29.878 [2024-11-20 13:52:36.753657] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:32:29.878 BaseBdev1 00:32:29.878 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:29.878 13:52:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@775 -- # sleep 1 00:32:30.814 13:52:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:32:30.814 13:52:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:30.814 13:52:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:30.814 13:52:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:30.814 13:52:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:30.814 13:52:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:30.814 13:52:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:30.814 13:52:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:30.814 13:52:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:30.814 13:52:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:30.814 13:52:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:30.814 13:52:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:30.814 13:52:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:30.814 13:52:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:30.814 13:52:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:30.814 13:52:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:30.814 "name": "raid_bdev1", 00:32:30.814 "uuid": "f0679e1a-4159-4f00-9c98-97c129b61d0e", 00:32:30.814 "strip_size_kb": 0, 00:32:30.814 "state": "online", 00:32:30.814 "raid_level": "raid1", 00:32:30.814 "superblock": true, 00:32:30.814 "num_base_bdevs": 4, 00:32:30.814 "num_base_bdevs_discovered": 2, 00:32:30.814 "num_base_bdevs_operational": 2, 00:32:30.814 "base_bdevs_list": [ 00:32:30.814 { 00:32:30.814 "name": null, 00:32:30.814 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:30.814 "is_configured": false, 00:32:30.814 "data_offset": 0, 00:32:30.814 "data_size": 63488 00:32:30.814 }, 00:32:30.814 { 00:32:30.814 "name": null, 00:32:30.814 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:30.814 "is_configured": false, 00:32:30.814 "data_offset": 2048, 00:32:30.814 "data_size": 63488 00:32:30.814 }, 00:32:30.814 { 00:32:30.814 "name": "BaseBdev3", 00:32:30.814 "uuid": "c5d06a40-515b-55af-b5ec-3ad678d14226", 00:32:30.814 "is_configured": true, 00:32:30.814 "data_offset": 2048, 00:32:30.814 "data_size": 63488 00:32:30.814 }, 00:32:30.814 { 00:32:30.814 "name": "BaseBdev4", 00:32:30.814 "uuid": "32609520-b04b-50c5-885a-8572452f2d36", 00:32:30.814 "is_configured": true, 00:32:30.814 "data_offset": 2048, 00:32:30.814 "data_size": 63488 00:32:30.814 } 00:32:30.814 ] 00:32:30.814 }' 00:32:30.814 13:52:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:30.814 13:52:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:31.072 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:32:31.072 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:31.072 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:32:31.072 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:32:31.072 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:31.072 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:31.072 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:31.072 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:31.072 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:31.072 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:31.072 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:31.072 "name": "raid_bdev1", 00:32:31.072 "uuid": "f0679e1a-4159-4f00-9c98-97c129b61d0e", 00:32:31.072 "strip_size_kb": 0, 00:32:31.072 "state": "online", 00:32:31.072 "raid_level": "raid1", 00:32:31.072 "superblock": true, 00:32:31.072 "num_base_bdevs": 4, 00:32:31.072 "num_base_bdevs_discovered": 2, 00:32:31.072 "num_base_bdevs_operational": 2, 00:32:31.072 "base_bdevs_list": [ 00:32:31.072 { 00:32:31.072 "name": null, 00:32:31.072 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:31.072 "is_configured": false, 00:32:31.072 "data_offset": 0, 00:32:31.072 "data_size": 63488 00:32:31.072 }, 00:32:31.072 { 00:32:31.072 "name": null, 00:32:31.072 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:31.072 "is_configured": false, 00:32:31.072 "data_offset": 2048, 00:32:31.072 "data_size": 63488 00:32:31.072 }, 00:32:31.072 { 00:32:31.072 "name": "BaseBdev3", 00:32:31.072 "uuid": "c5d06a40-515b-55af-b5ec-3ad678d14226", 00:32:31.072 "is_configured": true, 00:32:31.072 "data_offset": 2048, 00:32:31.072 "data_size": 63488 00:32:31.072 }, 00:32:31.072 { 00:32:31.072 "name": "BaseBdev4", 00:32:31.072 "uuid": "32609520-b04b-50c5-885a-8572452f2d36", 00:32:31.072 "is_configured": true, 00:32:31.072 "data_offset": 2048, 00:32:31.072 "data_size": 63488 00:32:31.072 } 00:32:31.072 ] 00:32:31.072 }' 00:32:31.072 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:31.330 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:32:31.330 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:31.330 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:32:31.330 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:32:31.330 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@652 -- # local es=0 00:32:31.330 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:32:31.330 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:32:31.330 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:32:31.330 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:32:31.330 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:32:31.330 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:32:31.330 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:31.330 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:31.330 [2024-11-20 13:52:38.193502] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:32:31.330 [2024-11-20 13:52:38.193667] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:32:31.330 [2024-11-20 13:52:38.193695] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:32:31.330 request: 00:32:31.330 { 00:32:31.330 "base_bdev": "BaseBdev1", 00:32:31.330 "raid_bdev": "raid_bdev1", 00:32:31.330 "method": "bdev_raid_add_base_bdev", 00:32:31.330 "req_id": 1 00:32:31.330 } 00:32:31.330 Got JSON-RPC error response 00:32:31.330 response: 00:32:31.330 { 00:32:31.330 "code": -22, 00:32:31.330 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:32:31.330 } 00:32:31.330 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:32:31.330 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # es=1 00:32:31.330 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:32:31.330 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:32:31.330 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:32:31.330 13:52:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@779 -- # sleep 1 00:32:32.277 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:32:32.277 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:32.277 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:32.277 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:32.277 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:32.277 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:32.277 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:32.277 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:32.277 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:32.277 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:32.277 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:32.277 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:32.277 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:32.277 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:32.277 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:32.277 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:32.278 "name": "raid_bdev1", 00:32:32.278 "uuid": "f0679e1a-4159-4f00-9c98-97c129b61d0e", 00:32:32.278 "strip_size_kb": 0, 00:32:32.278 "state": "online", 00:32:32.278 "raid_level": "raid1", 00:32:32.278 "superblock": true, 00:32:32.278 "num_base_bdevs": 4, 00:32:32.278 "num_base_bdevs_discovered": 2, 00:32:32.278 "num_base_bdevs_operational": 2, 00:32:32.278 "base_bdevs_list": [ 00:32:32.278 { 00:32:32.278 "name": null, 00:32:32.278 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:32.278 "is_configured": false, 00:32:32.278 "data_offset": 0, 00:32:32.278 "data_size": 63488 00:32:32.278 }, 00:32:32.278 { 00:32:32.278 "name": null, 00:32:32.278 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:32.278 "is_configured": false, 00:32:32.278 "data_offset": 2048, 00:32:32.278 "data_size": 63488 00:32:32.278 }, 00:32:32.278 { 00:32:32.278 "name": "BaseBdev3", 00:32:32.278 "uuid": "c5d06a40-515b-55af-b5ec-3ad678d14226", 00:32:32.278 "is_configured": true, 00:32:32.278 "data_offset": 2048, 00:32:32.278 "data_size": 63488 00:32:32.278 }, 00:32:32.278 { 00:32:32.278 "name": "BaseBdev4", 00:32:32.278 "uuid": "32609520-b04b-50c5-885a-8572452f2d36", 00:32:32.278 "is_configured": true, 00:32:32.278 "data_offset": 2048, 00:32:32.278 "data_size": 63488 00:32:32.278 } 00:32:32.278 ] 00:32:32.278 }' 00:32:32.278 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:32.278 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:32.542 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:32:32.542 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:32.542 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:32:32.542 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:32:32.542 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:32.543 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:32.543 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:32.543 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:32.543 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:32.543 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:32.543 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:32.543 "name": "raid_bdev1", 00:32:32.543 "uuid": "f0679e1a-4159-4f00-9c98-97c129b61d0e", 00:32:32.543 "strip_size_kb": 0, 00:32:32.543 "state": "online", 00:32:32.543 "raid_level": "raid1", 00:32:32.543 "superblock": true, 00:32:32.543 "num_base_bdevs": 4, 00:32:32.543 "num_base_bdevs_discovered": 2, 00:32:32.543 "num_base_bdevs_operational": 2, 00:32:32.543 "base_bdevs_list": [ 00:32:32.543 { 00:32:32.543 "name": null, 00:32:32.543 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:32.543 "is_configured": false, 00:32:32.543 "data_offset": 0, 00:32:32.543 "data_size": 63488 00:32:32.543 }, 00:32:32.543 { 00:32:32.543 "name": null, 00:32:32.543 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:32.543 "is_configured": false, 00:32:32.543 "data_offset": 2048, 00:32:32.543 "data_size": 63488 00:32:32.543 }, 00:32:32.543 { 00:32:32.543 "name": "BaseBdev3", 00:32:32.543 "uuid": "c5d06a40-515b-55af-b5ec-3ad678d14226", 00:32:32.543 "is_configured": true, 00:32:32.543 "data_offset": 2048, 00:32:32.543 "data_size": 63488 00:32:32.543 }, 00:32:32.543 { 00:32:32.543 "name": "BaseBdev4", 00:32:32.543 "uuid": "32609520-b04b-50c5-885a-8572452f2d36", 00:32:32.543 "is_configured": true, 00:32:32.543 "data_offset": 2048, 00:32:32.543 "data_size": 63488 00:32:32.543 } 00:32:32.543 ] 00:32:32.543 }' 00:32:32.543 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:32.543 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:32:32.543 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:32.801 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:32:32.801 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@784 -- # killprocess 76938 00:32:32.801 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@954 -- # '[' -z 76938 ']' 00:32:32.801 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@958 -- # kill -0 76938 00:32:32.801 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # uname 00:32:32.801 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:32:32.801 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 76938 00:32:32.801 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:32:32.801 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:32:32.801 killing process with pid 76938 00:32:32.801 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 76938' 00:32:32.801 Received shutdown signal, test time was about 16.090127 seconds 00:32:32.801 00:32:32.801 Latency(us) 00:32:32.801 [2024-11-20T13:52:39.860Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:32.801 [2024-11-20T13:52:39.860Z] =================================================================================================================== 00:32:32.801 [2024-11-20T13:52:39.860Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:32:32.801 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@973 -- # kill 76938 00:32:32.801 [2024-11-20 13:52:39.644099] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:32:32.801 13:52:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@978 -- # wait 76938 00:32:32.801 [2024-11-20 13:52:39.644229] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:32:32.801 [2024-11-20 13:52:39.644302] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:32:32.801 [2024-11-20 13:52:39.644320] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:32:33.059 [2024-11-20 13:52:39.861126] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:32:33.627 13:52:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@786 -- # return 0 00:32:33.627 00:32:33.627 real 0m18.544s 00:32:33.627 user 0m23.531s 00:32:33.627 sys 0m1.961s 00:32:33.627 ************************************ 00:32:33.627 END TEST raid_rebuild_test_sb_io 00:32:33.627 ************************************ 00:32:33.627 13:52:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:32:33.627 13:52:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:32:33.627 13:52:40 bdev_raid -- bdev/bdev_raid.sh@985 -- # for n in {3..4} 00:32:33.627 13:52:40 bdev_raid -- bdev/bdev_raid.sh@986 -- # run_test raid5f_state_function_test raid_state_function_test raid5f 3 false 00:32:33.627 13:52:40 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:32:33.627 13:52:40 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:32:33.627 13:52:40 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:32:33.627 ************************************ 00:32:33.627 START TEST raid5f_state_function_test 00:32:33.627 ************************************ 00:32:33.627 13:52:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 3 false 00:32:33.627 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:32:33.627 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:32:33.627 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:32:33.627 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=77637 00:32:33.628 Process raid pid: 77637 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 77637' 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 77637 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 77637 ']' 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:32:33.628 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:32:33.628 13:52:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:33.628 [2024-11-20 13:52:40.628547] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:32:33.628 [2024-11-20 13:52:40.628663] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:32:33.889 [2024-11-20 13:52:40.788385] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:33.889 [2024-11-20 13:52:40.907388] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:32:34.147 [2024-11-20 13:52:41.056899] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:32:34.147 [2024-11-20 13:52:41.056953] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:32:34.717 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:32:34.717 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:32:34.717 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:32:34.717 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:34.717 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:34.717 [2024-11-20 13:52:41.519124] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:32:34.717 [2024-11-20 13:52:41.519185] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:32:34.717 [2024-11-20 13:52:41.519195] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:32:34.717 [2024-11-20 13:52:41.519205] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:32:34.717 [2024-11-20 13:52:41.519212] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:32:34.717 [2024-11-20 13:52:41.519220] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:32:34.717 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:34.717 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:32:34.717 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:34.717 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:34.717 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:34.717 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:34.717 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:34.717 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:34.717 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:34.717 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:34.717 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:34.717 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:34.717 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:34.717 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:34.717 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:34.717 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:34.717 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:34.717 "name": "Existed_Raid", 00:32:34.717 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:34.717 "strip_size_kb": 64, 00:32:34.717 "state": "configuring", 00:32:34.717 "raid_level": "raid5f", 00:32:34.717 "superblock": false, 00:32:34.717 "num_base_bdevs": 3, 00:32:34.717 "num_base_bdevs_discovered": 0, 00:32:34.717 "num_base_bdevs_operational": 3, 00:32:34.717 "base_bdevs_list": [ 00:32:34.717 { 00:32:34.717 "name": "BaseBdev1", 00:32:34.717 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:34.717 "is_configured": false, 00:32:34.717 "data_offset": 0, 00:32:34.717 "data_size": 0 00:32:34.717 }, 00:32:34.717 { 00:32:34.717 "name": "BaseBdev2", 00:32:34.717 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:34.717 "is_configured": false, 00:32:34.717 "data_offset": 0, 00:32:34.717 "data_size": 0 00:32:34.717 }, 00:32:34.717 { 00:32:34.717 "name": "BaseBdev3", 00:32:34.718 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:34.718 "is_configured": false, 00:32:34.718 "data_offset": 0, 00:32:34.718 "data_size": 0 00:32:34.718 } 00:32:34.718 ] 00:32:34.718 }' 00:32:34.718 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:34.718 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:34.979 [2024-11-20 13:52:41.847145] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:32:34.979 [2024-11-20 13:52:41.847184] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:34.979 [2024-11-20 13:52:41.855142] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:32:34.979 [2024-11-20 13:52:41.855185] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:32:34.979 [2024-11-20 13:52:41.855194] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:32:34.979 [2024-11-20 13:52:41.855203] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:32:34.979 [2024-11-20 13:52:41.855209] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:32:34.979 [2024-11-20 13:52:41.855218] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:34.979 [2024-11-20 13:52:41.890043] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:32:34.979 BaseBdev1 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:34.979 [ 00:32:34.979 { 00:32:34.979 "name": "BaseBdev1", 00:32:34.979 "aliases": [ 00:32:34.979 "a13cf5ae-4177-40ca-aa5c-d62cf9852654" 00:32:34.979 ], 00:32:34.979 "product_name": "Malloc disk", 00:32:34.979 "block_size": 512, 00:32:34.979 "num_blocks": 65536, 00:32:34.979 "uuid": "a13cf5ae-4177-40ca-aa5c-d62cf9852654", 00:32:34.979 "assigned_rate_limits": { 00:32:34.979 "rw_ios_per_sec": 0, 00:32:34.979 "rw_mbytes_per_sec": 0, 00:32:34.979 "r_mbytes_per_sec": 0, 00:32:34.979 "w_mbytes_per_sec": 0 00:32:34.979 }, 00:32:34.979 "claimed": true, 00:32:34.979 "claim_type": "exclusive_write", 00:32:34.979 "zoned": false, 00:32:34.979 "supported_io_types": { 00:32:34.979 "read": true, 00:32:34.979 "write": true, 00:32:34.979 "unmap": true, 00:32:34.979 "flush": true, 00:32:34.979 "reset": true, 00:32:34.979 "nvme_admin": false, 00:32:34.979 "nvme_io": false, 00:32:34.979 "nvme_io_md": false, 00:32:34.979 "write_zeroes": true, 00:32:34.979 "zcopy": true, 00:32:34.979 "get_zone_info": false, 00:32:34.979 "zone_management": false, 00:32:34.979 "zone_append": false, 00:32:34.979 "compare": false, 00:32:34.979 "compare_and_write": false, 00:32:34.979 "abort": true, 00:32:34.979 "seek_hole": false, 00:32:34.979 "seek_data": false, 00:32:34.979 "copy": true, 00:32:34.979 "nvme_iov_md": false 00:32:34.979 }, 00:32:34.979 "memory_domains": [ 00:32:34.979 { 00:32:34.979 "dma_device_id": "system", 00:32:34.979 "dma_device_type": 1 00:32:34.979 }, 00:32:34.979 { 00:32:34.979 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:32:34.979 "dma_device_type": 2 00:32:34.979 } 00:32:34.979 ], 00:32:34.979 "driver_specific": {} 00:32:34.979 } 00:32:34.979 ] 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:34.979 "name": "Existed_Raid", 00:32:34.979 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:34.979 "strip_size_kb": 64, 00:32:34.979 "state": "configuring", 00:32:34.979 "raid_level": "raid5f", 00:32:34.979 "superblock": false, 00:32:34.979 "num_base_bdevs": 3, 00:32:34.979 "num_base_bdevs_discovered": 1, 00:32:34.979 "num_base_bdevs_operational": 3, 00:32:34.979 "base_bdevs_list": [ 00:32:34.979 { 00:32:34.979 "name": "BaseBdev1", 00:32:34.979 "uuid": "a13cf5ae-4177-40ca-aa5c-d62cf9852654", 00:32:34.979 "is_configured": true, 00:32:34.979 "data_offset": 0, 00:32:34.979 "data_size": 65536 00:32:34.979 }, 00:32:34.979 { 00:32:34.979 "name": "BaseBdev2", 00:32:34.979 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:34.979 "is_configured": false, 00:32:34.979 "data_offset": 0, 00:32:34.979 "data_size": 0 00:32:34.979 }, 00:32:34.979 { 00:32:34.979 "name": "BaseBdev3", 00:32:34.979 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:34.979 "is_configured": false, 00:32:34.979 "data_offset": 0, 00:32:34.979 "data_size": 0 00:32:34.979 } 00:32:34.979 ] 00:32:34.979 }' 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:34.979 13:52:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:35.238 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:32:35.238 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:35.239 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:35.239 [2024-11-20 13:52:42.242163] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:32:35.239 [2024-11-20 13:52:42.242219] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:32:35.239 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:35.239 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:32:35.239 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:35.239 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:35.239 [2024-11-20 13:52:42.250208] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:32:35.239 [2024-11-20 13:52:42.252182] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:32:35.239 [2024-11-20 13:52:42.252222] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:32:35.239 [2024-11-20 13:52:42.252233] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:32:35.239 [2024-11-20 13:52:42.252242] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:32:35.239 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:35.239 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:32:35.239 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:32:35.239 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:32:35.239 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:35.239 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:35.239 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:35.239 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:35.239 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:35.239 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:35.239 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:35.239 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:35.239 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:35.239 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:35.239 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:35.239 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:35.239 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:35.239 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:35.239 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:35.239 "name": "Existed_Raid", 00:32:35.239 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:35.239 "strip_size_kb": 64, 00:32:35.239 "state": "configuring", 00:32:35.239 "raid_level": "raid5f", 00:32:35.239 "superblock": false, 00:32:35.239 "num_base_bdevs": 3, 00:32:35.239 "num_base_bdevs_discovered": 1, 00:32:35.239 "num_base_bdevs_operational": 3, 00:32:35.239 "base_bdevs_list": [ 00:32:35.239 { 00:32:35.239 "name": "BaseBdev1", 00:32:35.239 "uuid": "a13cf5ae-4177-40ca-aa5c-d62cf9852654", 00:32:35.239 "is_configured": true, 00:32:35.239 "data_offset": 0, 00:32:35.239 "data_size": 65536 00:32:35.239 }, 00:32:35.239 { 00:32:35.239 "name": "BaseBdev2", 00:32:35.239 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:35.239 "is_configured": false, 00:32:35.239 "data_offset": 0, 00:32:35.239 "data_size": 0 00:32:35.239 }, 00:32:35.239 { 00:32:35.239 "name": "BaseBdev3", 00:32:35.239 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:35.239 "is_configured": false, 00:32:35.239 "data_offset": 0, 00:32:35.239 "data_size": 0 00:32:35.239 } 00:32:35.239 ] 00:32:35.239 }' 00:32:35.239 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:35.239 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:35.570 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:32:35.570 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:35.570 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:35.570 [2024-11-20 13:52:42.595496] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:32:35.570 BaseBdev2 00:32:35.570 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:35.570 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:32:35.570 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:32:35.570 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:32:35.570 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:32:35.570 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:32:35.570 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:32:35.570 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:32:35.570 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:35.570 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:35.570 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:35.570 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:32:35.571 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:35.571 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:35.873 [ 00:32:35.873 { 00:32:35.873 "name": "BaseBdev2", 00:32:35.873 "aliases": [ 00:32:35.873 "7cb44ad1-0cf3-41e3-a47a-801fe94270dc" 00:32:35.873 ], 00:32:35.873 "product_name": "Malloc disk", 00:32:35.873 "block_size": 512, 00:32:35.873 "num_blocks": 65536, 00:32:35.873 "uuid": "7cb44ad1-0cf3-41e3-a47a-801fe94270dc", 00:32:35.873 "assigned_rate_limits": { 00:32:35.873 "rw_ios_per_sec": 0, 00:32:35.873 "rw_mbytes_per_sec": 0, 00:32:35.873 "r_mbytes_per_sec": 0, 00:32:35.873 "w_mbytes_per_sec": 0 00:32:35.873 }, 00:32:35.873 "claimed": true, 00:32:35.873 "claim_type": "exclusive_write", 00:32:35.873 "zoned": false, 00:32:35.873 "supported_io_types": { 00:32:35.873 "read": true, 00:32:35.873 "write": true, 00:32:35.873 "unmap": true, 00:32:35.873 "flush": true, 00:32:35.873 "reset": true, 00:32:35.873 "nvme_admin": false, 00:32:35.873 "nvme_io": false, 00:32:35.873 "nvme_io_md": false, 00:32:35.873 "write_zeroes": true, 00:32:35.873 "zcopy": true, 00:32:35.873 "get_zone_info": false, 00:32:35.873 "zone_management": false, 00:32:35.873 "zone_append": false, 00:32:35.873 "compare": false, 00:32:35.873 "compare_and_write": false, 00:32:35.873 "abort": true, 00:32:35.873 "seek_hole": false, 00:32:35.873 "seek_data": false, 00:32:35.873 "copy": true, 00:32:35.873 "nvme_iov_md": false 00:32:35.873 }, 00:32:35.873 "memory_domains": [ 00:32:35.873 { 00:32:35.873 "dma_device_id": "system", 00:32:35.873 "dma_device_type": 1 00:32:35.873 }, 00:32:35.873 { 00:32:35.873 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:32:35.873 "dma_device_type": 2 00:32:35.873 } 00:32:35.873 ], 00:32:35.873 "driver_specific": {} 00:32:35.873 } 00:32:35.873 ] 00:32:35.873 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:35.873 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:32:35.873 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:32:35.873 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:32:35.873 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:32:35.873 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:35.873 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:35.873 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:35.873 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:35.873 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:35.873 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:35.873 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:35.873 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:35.873 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:35.873 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:35.873 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:35.873 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:35.873 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:35.873 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:35.873 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:35.873 "name": "Existed_Raid", 00:32:35.873 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:35.873 "strip_size_kb": 64, 00:32:35.873 "state": "configuring", 00:32:35.873 "raid_level": "raid5f", 00:32:35.873 "superblock": false, 00:32:35.873 "num_base_bdevs": 3, 00:32:35.873 "num_base_bdevs_discovered": 2, 00:32:35.873 "num_base_bdevs_operational": 3, 00:32:35.873 "base_bdevs_list": [ 00:32:35.873 { 00:32:35.873 "name": "BaseBdev1", 00:32:35.873 "uuid": "a13cf5ae-4177-40ca-aa5c-d62cf9852654", 00:32:35.873 "is_configured": true, 00:32:35.873 "data_offset": 0, 00:32:35.873 "data_size": 65536 00:32:35.873 }, 00:32:35.873 { 00:32:35.873 "name": "BaseBdev2", 00:32:35.873 "uuid": "7cb44ad1-0cf3-41e3-a47a-801fe94270dc", 00:32:35.873 "is_configured": true, 00:32:35.873 "data_offset": 0, 00:32:35.873 "data_size": 65536 00:32:35.873 }, 00:32:35.873 { 00:32:35.873 "name": "BaseBdev3", 00:32:35.873 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:35.873 "is_configured": false, 00:32:35.873 "data_offset": 0, 00:32:35.873 "data_size": 0 00:32:35.873 } 00:32:35.873 ] 00:32:35.873 }' 00:32:35.873 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:35.873 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:35.873 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:32:35.873 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:35.873 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:36.132 [2024-11-20 13:52:42.971314] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:32:36.132 [2024-11-20 13:52:42.971374] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:32:36.132 [2024-11-20 13:52:42.971390] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:32:36.132 [2024-11-20 13:52:42.971657] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:32:36.132 [2024-11-20 13:52:42.975474] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:32:36.132 [2024-11-20 13:52:42.975497] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:32:36.132 [2024-11-20 13:52:42.975783] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:36.132 BaseBdev3 00:32:36.132 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:36.132 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:32:36.132 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:32:36.132 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:32:36.132 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:32:36.132 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:32:36.132 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:32:36.132 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:32:36.132 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:36.132 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:36.132 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:36.132 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:32:36.132 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:36.132 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:36.132 [ 00:32:36.132 { 00:32:36.132 "name": "BaseBdev3", 00:32:36.132 "aliases": [ 00:32:36.132 "99c8f382-9435-4eae-b00c-8806cd9665b2" 00:32:36.132 ], 00:32:36.132 "product_name": "Malloc disk", 00:32:36.132 "block_size": 512, 00:32:36.132 "num_blocks": 65536, 00:32:36.132 "uuid": "99c8f382-9435-4eae-b00c-8806cd9665b2", 00:32:36.132 "assigned_rate_limits": { 00:32:36.132 "rw_ios_per_sec": 0, 00:32:36.132 "rw_mbytes_per_sec": 0, 00:32:36.132 "r_mbytes_per_sec": 0, 00:32:36.132 "w_mbytes_per_sec": 0 00:32:36.132 }, 00:32:36.132 "claimed": true, 00:32:36.132 "claim_type": "exclusive_write", 00:32:36.132 "zoned": false, 00:32:36.132 "supported_io_types": { 00:32:36.132 "read": true, 00:32:36.132 "write": true, 00:32:36.132 "unmap": true, 00:32:36.132 "flush": true, 00:32:36.132 "reset": true, 00:32:36.132 "nvme_admin": false, 00:32:36.132 "nvme_io": false, 00:32:36.132 "nvme_io_md": false, 00:32:36.132 "write_zeroes": true, 00:32:36.133 "zcopy": true, 00:32:36.133 "get_zone_info": false, 00:32:36.133 "zone_management": false, 00:32:36.133 "zone_append": false, 00:32:36.133 "compare": false, 00:32:36.133 "compare_and_write": false, 00:32:36.133 "abort": true, 00:32:36.133 "seek_hole": false, 00:32:36.133 "seek_data": false, 00:32:36.133 "copy": true, 00:32:36.133 "nvme_iov_md": false 00:32:36.133 }, 00:32:36.133 "memory_domains": [ 00:32:36.133 { 00:32:36.133 "dma_device_id": "system", 00:32:36.133 "dma_device_type": 1 00:32:36.133 }, 00:32:36.133 { 00:32:36.133 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:32:36.133 "dma_device_type": 2 00:32:36.133 } 00:32:36.133 ], 00:32:36.133 "driver_specific": {} 00:32:36.133 } 00:32:36.133 ] 00:32:36.133 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:36.133 13:52:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:32:36.133 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:32:36.133 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:32:36.133 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:32:36.133 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:36.133 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:36.133 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:36.133 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:36.133 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:36.133 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:36.133 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:36.133 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:36.133 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:36.133 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:36.133 13:52:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:36.133 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:36.133 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:36.133 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:36.133 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:36.133 "name": "Existed_Raid", 00:32:36.133 "uuid": "122764ee-6d5a-4044-91b0-403fb840e2da", 00:32:36.133 "strip_size_kb": 64, 00:32:36.133 "state": "online", 00:32:36.133 "raid_level": "raid5f", 00:32:36.133 "superblock": false, 00:32:36.133 "num_base_bdevs": 3, 00:32:36.133 "num_base_bdevs_discovered": 3, 00:32:36.133 "num_base_bdevs_operational": 3, 00:32:36.133 "base_bdevs_list": [ 00:32:36.133 { 00:32:36.133 "name": "BaseBdev1", 00:32:36.133 "uuid": "a13cf5ae-4177-40ca-aa5c-d62cf9852654", 00:32:36.133 "is_configured": true, 00:32:36.133 "data_offset": 0, 00:32:36.133 "data_size": 65536 00:32:36.133 }, 00:32:36.133 { 00:32:36.133 "name": "BaseBdev2", 00:32:36.133 "uuid": "7cb44ad1-0cf3-41e3-a47a-801fe94270dc", 00:32:36.133 "is_configured": true, 00:32:36.133 "data_offset": 0, 00:32:36.133 "data_size": 65536 00:32:36.133 }, 00:32:36.133 { 00:32:36.133 "name": "BaseBdev3", 00:32:36.133 "uuid": "99c8f382-9435-4eae-b00c-8806cd9665b2", 00:32:36.133 "is_configured": true, 00:32:36.133 "data_offset": 0, 00:32:36.133 "data_size": 65536 00:32:36.133 } 00:32:36.133 ] 00:32:36.133 }' 00:32:36.133 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:36.133 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:36.391 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:32:36.391 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:32:36.391 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:32:36.391 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:32:36.391 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:32:36.391 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:32:36.391 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:32:36.391 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:32:36.391 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:36.391 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:36.391 [2024-11-20 13:52:43.304423] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:32:36.391 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:36.391 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:32:36.391 "name": "Existed_Raid", 00:32:36.391 "aliases": [ 00:32:36.391 "122764ee-6d5a-4044-91b0-403fb840e2da" 00:32:36.391 ], 00:32:36.391 "product_name": "Raid Volume", 00:32:36.391 "block_size": 512, 00:32:36.391 "num_blocks": 131072, 00:32:36.391 "uuid": "122764ee-6d5a-4044-91b0-403fb840e2da", 00:32:36.391 "assigned_rate_limits": { 00:32:36.391 "rw_ios_per_sec": 0, 00:32:36.391 "rw_mbytes_per_sec": 0, 00:32:36.391 "r_mbytes_per_sec": 0, 00:32:36.391 "w_mbytes_per_sec": 0 00:32:36.391 }, 00:32:36.391 "claimed": false, 00:32:36.391 "zoned": false, 00:32:36.391 "supported_io_types": { 00:32:36.391 "read": true, 00:32:36.391 "write": true, 00:32:36.391 "unmap": false, 00:32:36.391 "flush": false, 00:32:36.391 "reset": true, 00:32:36.391 "nvme_admin": false, 00:32:36.391 "nvme_io": false, 00:32:36.391 "nvme_io_md": false, 00:32:36.391 "write_zeroes": true, 00:32:36.391 "zcopy": false, 00:32:36.391 "get_zone_info": false, 00:32:36.391 "zone_management": false, 00:32:36.391 "zone_append": false, 00:32:36.391 "compare": false, 00:32:36.391 "compare_and_write": false, 00:32:36.391 "abort": false, 00:32:36.391 "seek_hole": false, 00:32:36.391 "seek_data": false, 00:32:36.391 "copy": false, 00:32:36.391 "nvme_iov_md": false 00:32:36.391 }, 00:32:36.391 "driver_specific": { 00:32:36.391 "raid": { 00:32:36.391 "uuid": "122764ee-6d5a-4044-91b0-403fb840e2da", 00:32:36.391 "strip_size_kb": 64, 00:32:36.391 "state": "online", 00:32:36.391 "raid_level": "raid5f", 00:32:36.391 "superblock": false, 00:32:36.391 "num_base_bdevs": 3, 00:32:36.391 "num_base_bdevs_discovered": 3, 00:32:36.391 "num_base_bdevs_operational": 3, 00:32:36.391 "base_bdevs_list": [ 00:32:36.391 { 00:32:36.391 "name": "BaseBdev1", 00:32:36.391 "uuid": "a13cf5ae-4177-40ca-aa5c-d62cf9852654", 00:32:36.391 "is_configured": true, 00:32:36.391 "data_offset": 0, 00:32:36.391 "data_size": 65536 00:32:36.391 }, 00:32:36.391 { 00:32:36.392 "name": "BaseBdev2", 00:32:36.392 "uuid": "7cb44ad1-0cf3-41e3-a47a-801fe94270dc", 00:32:36.392 "is_configured": true, 00:32:36.392 "data_offset": 0, 00:32:36.392 "data_size": 65536 00:32:36.392 }, 00:32:36.392 { 00:32:36.392 "name": "BaseBdev3", 00:32:36.392 "uuid": "99c8f382-9435-4eae-b00c-8806cd9665b2", 00:32:36.392 "is_configured": true, 00:32:36.392 "data_offset": 0, 00:32:36.392 "data_size": 65536 00:32:36.392 } 00:32:36.392 ] 00:32:36.392 } 00:32:36.392 } 00:32:36.392 }' 00:32:36.392 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:32:36.392 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:32:36.392 BaseBdev2 00:32:36.392 BaseBdev3' 00:32:36.392 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:36.392 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:32:36.392 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:32:36.392 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:36.392 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:32:36.392 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:36.392 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:36.392 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:36.392 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:32:36.392 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:32:36.392 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:32:36.392 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:36.392 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:32:36.392 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:36.392 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:36.651 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:36.651 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:32:36.651 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:32:36.651 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:32:36.651 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:36.651 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:32:36.651 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:36.651 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:36.651 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:36.651 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:32:36.651 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:32:36.651 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:32:36.651 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:36.651 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:36.651 [2024-11-20 13:52:43.516305] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:32:36.651 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:36.651 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:32:36.651 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:32:36.651 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:32:36.652 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:32:36.652 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:32:36.652 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 2 00:32:36.652 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:36.652 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:36.652 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:36.652 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:36.652 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:36.652 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:36.652 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:36.652 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:36.652 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:36.652 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:36.652 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:36.652 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:36.652 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:36.652 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:36.652 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:36.652 "name": "Existed_Raid", 00:32:36.652 "uuid": "122764ee-6d5a-4044-91b0-403fb840e2da", 00:32:36.652 "strip_size_kb": 64, 00:32:36.652 "state": "online", 00:32:36.652 "raid_level": "raid5f", 00:32:36.652 "superblock": false, 00:32:36.652 "num_base_bdevs": 3, 00:32:36.652 "num_base_bdevs_discovered": 2, 00:32:36.652 "num_base_bdevs_operational": 2, 00:32:36.652 "base_bdevs_list": [ 00:32:36.652 { 00:32:36.652 "name": null, 00:32:36.652 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:36.652 "is_configured": false, 00:32:36.652 "data_offset": 0, 00:32:36.652 "data_size": 65536 00:32:36.652 }, 00:32:36.652 { 00:32:36.652 "name": "BaseBdev2", 00:32:36.652 "uuid": "7cb44ad1-0cf3-41e3-a47a-801fe94270dc", 00:32:36.652 "is_configured": true, 00:32:36.652 "data_offset": 0, 00:32:36.652 "data_size": 65536 00:32:36.652 }, 00:32:36.652 { 00:32:36.652 "name": "BaseBdev3", 00:32:36.652 "uuid": "99c8f382-9435-4eae-b00c-8806cd9665b2", 00:32:36.652 "is_configured": true, 00:32:36.652 "data_offset": 0, 00:32:36.652 "data_size": 65536 00:32:36.652 } 00:32:36.652 ] 00:32:36.652 }' 00:32:36.652 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:36.652 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:36.913 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:32:36.913 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:32:36.913 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:32:36.913 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:36.913 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:36.913 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:36.913 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:36.913 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:32:36.913 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:32:36.913 13:52:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:32:36.913 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:36.913 13:52:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:36.913 [2024-11-20 13:52:43.964319] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:32:36.913 [2024-11-20 13:52:43.964429] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:32:37.174 [2024-11-20 13:52:44.027878] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:32:37.174 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:37.174 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:32:37.174 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:32:37.174 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:37.174 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:37.174 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:37.174 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:32:37.174 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:37.174 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:32:37.174 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:32:37.174 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:32:37.174 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:37.174 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:37.174 [2024-11-20 13:52:44.071938] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:32:37.175 [2024-11-20 13:52:44.072088] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:37.175 BaseBdev2 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:37.175 [ 00:32:37.175 { 00:32:37.175 "name": "BaseBdev2", 00:32:37.175 "aliases": [ 00:32:37.175 "dfa6467f-25ae-44e2-877c-ff7200bf1783" 00:32:37.175 ], 00:32:37.175 "product_name": "Malloc disk", 00:32:37.175 "block_size": 512, 00:32:37.175 "num_blocks": 65536, 00:32:37.175 "uuid": "dfa6467f-25ae-44e2-877c-ff7200bf1783", 00:32:37.175 "assigned_rate_limits": { 00:32:37.175 "rw_ios_per_sec": 0, 00:32:37.175 "rw_mbytes_per_sec": 0, 00:32:37.175 "r_mbytes_per_sec": 0, 00:32:37.175 "w_mbytes_per_sec": 0 00:32:37.175 }, 00:32:37.175 "claimed": false, 00:32:37.175 "zoned": false, 00:32:37.175 "supported_io_types": { 00:32:37.175 "read": true, 00:32:37.175 "write": true, 00:32:37.175 "unmap": true, 00:32:37.175 "flush": true, 00:32:37.175 "reset": true, 00:32:37.175 "nvme_admin": false, 00:32:37.175 "nvme_io": false, 00:32:37.175 "nvme_io_md": false, 00:32:37.175 "write_zeroes": true, 00:32:37.175 "zcopy": true, 00:32:37.175 "get_zone_info": false, 00:32:37.175 "zone_management": false, 00:32:37.175 "zone_append": false, 00:32:37.175 "compare": false, 00:32:37.175 "compare_and_write": false, 00:32:37.175 "abort": true, 00:32:37.175 "seek_hole": false, 00:32:37.175 "seek_data": false, 00:32:37.175 "copy": true, 00:32:37.175 "nvme_iov_md": false 00:32:37.175 }, 00:32:37.175 "memory_domains": [ 00:32:37.175 { 00:32:37.175 "dma_device_id": "system", 00:32:37.175 "dma_device_type": 1 00:32:37.175 }, 00:32:37.175 { 00:32:37.175 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:32:37.175 "dma_device_type": 2 00:32:37.175 } 00:32:37.175 ], 00:32:37.175 "driver_specific": {} 00:32:37.175 } 00:32:37.175 ] 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:37.175 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:37.437 BaseBdev3 00:32:37.437 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:37.437 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:32:37.437 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:32:37.437 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:32:37.437 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:32:37.437 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:32:37.437 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:32:37.437 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:32:37.437 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:37.437 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:37.437 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:37.437 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:32:37.437 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:37.437 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:37.437 [ 00:32:37.437 { 00:32:37.437 "name": "BaseBdev3", 00:32:37.437 "aliases": [ 00:32:37.437 "6a0e8948-ca88-432c-9f83-8b3698a2e7be" 00:32:37.437 ], 00:32:37.437 "product_name": "Malloc disk", 00:32:37.437 "block_size": 512, 00:32:37.437 "num_blocks": 65536, 00:32:37.437 "uuid": "6a0e8948-ca88-432c-9f83-8b3698a2e7be", 00:32:37.437 "assigned_rate_limits": { 00:32:37.437 "rw_ios_per_sec": 0, 00:32:37.437 "rw_mbytes_per_sec": 0, 00:32:37.437 "r_mbytes_per_sec": 0, 00:32:37.437 "w_mbytes_per_sec": 0 00:32:37.437 }, 00:32:37.437 "claimed": false, 00:32:37.437 "zoned": false, 00:32:37.437 "supported_io_types": { 00:32:37.437 "read": true, 00:32:37.437 "write": true, 00:32:37.437 "unmap": true, 00:32:37.437 "flush": true, 00:32:37.437 "reset": true, 00:32:37.437 "nvme_admin": false, 00:32:37.437 "nvme_io": false, 00:32:37.437 "nvme_io_md": false, 00:32:37.438 "write_zeroes": true, 00:32:37.438 "zcopy": true, 00:32:37.438 "get_zone_info": false, 00:32:37.438 "zone_management": false, 00:32:37.438 "zone_append": false, 00:32:37.438 "compare": false, 00:32:37.438 "compare_and_write": false, 00:32:37.438 "abort": true, 00:32:37.438 "seek_hole": false, 00:32:37.438 "seek_data": false, 00:32:37.438 "copy": true, 00:32:37.438 "nvme_iov_md": false 00:32:37.438 }, 00:32:37.438 "memory_domains": [ 00:32:37.438 { 00:32:37.438 "dma_device_id": "system", 00:32:37.438 "dma_device_type": 1 00:32:37.438 }, 00:32:37.438 { 00:32:37.438 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:32:37.438 "dma_device_type": 2 00:32:37.438 } 00:32:37.438 ], 00:32:37.438 "driver_specific": {} 00:32:37.438 } 00:32:37.438 ] 00:32:37.438 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:37.438 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:32:37.438 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:32:37.438 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:32:37.438 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:32:37.438 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:37.438 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:37.438 [2024-11-20 13:52:44.288147] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:32:37.438 [2024-11-20 13:52:44.288192] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:32:37.438 [2024-11-20 13:52:44.288213] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:32:37.438 [2024-11-20 13:52:44.290138] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:32:37.438 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:37.438 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:32:37.438 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:37.438 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:37.438 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:37.438 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:37.438 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:37.438 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:37.438 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:37.438 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:37.438 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:37.438 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:37.438 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:37.438 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:37.438 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:37.438 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:37.438 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:37.438 "name": "Existed_Raid", 00:32:37.438 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:37.438 "strip_size_kb": 64, 00:32:37.438 "state": "configuring", 00:32:37.438 "raid_level": "raid5f", 00:32:37.438 "superblock": false, 00:32:37.438 "num_base_bdevs": 3, 00:32:37.438 "num_base_bdevs_discovered": 2, 00:32:37.438 "num_base_bdevs_operational": 3, 00:32:37.438 "base_bdevs_list": [ 00:32:37.438 { 00:32:37.438 "name": "BaseBdev1", 00:32:37.438 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:37.438 "is_configured": false, 00:32:37.438 "data_offset": 0, 00:32:37.438 "data_size": 0 00:32:37.438 }, 00:32:37.438 { 00:32:37.438 "name": "BaseBdev2", 00:32:37.438 "uuid": "dfa6467f-25ae-44e2-877c-ff7200bf1783", 00:32:37.438 "is_configured": true, 00:32:37.438 "data_offset": 0, 00:32:37.438 "data_size": 65536 00:32:37.438 }, 00:32:37.438 { 00:32:37.438 "name": "BaseBdev3", 00:32:37.438 "uuid": "6a0e8948-ca88-432c-9f83-8b3698a2e7be", 00:32:37.438 "is_configured": true, 00:32:37.438 "data_offset": 0, 00:32:37.438 "data_size": 65536 00:32:37.438 } 00:32:37.438 ] 00:32:37.438 }' 00:32:37.438 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:37.438 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:37.699 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:32:37.699 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:37.699 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:37.699 [2024-11-20 13:52:44.624256] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:32:37.699 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:37.699 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:32:37.699 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:37.699 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:37.699 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:37.700 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:37.700 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:37.700 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:37.700 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:37.700 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:37.700 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:37.700 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:37.700 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:37.700 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:37.700 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:37.700 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:37.700 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:37.700 "name": "Existed_Raid", 00:32:37.700 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:37.700 "strip_size_kb": 64, 00:32:37.700 "state": "configuring", 00:32:37.700 "raid_level": "raid5f", 00:32:37.700 "superblock": false, 00:32:37.700 "num_base_bdevs": 3, 00:32:37.700 "num_base_bdevs_discovered": 1, 00:32:37.700 "num_base_bdevs_operational": 3, 00:32:37.700 "base_bdevs_list": [ 00:32:37.700 { 00:32:37.700 "name": "BaseBdev1", 00:32:37.700 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:37.700 "is_configured": false, 00:32:37.700 "data_offset": 0, 00:32:37.700 "data_size": 0 00:32:37.700 }, 00:32:37.700 { 00:32:37.700 "name": null, 00:32:37.700 "uuid": "dfa6467f-25ae-44e2-877c-ff7200bf1783", 00:32:37.700 "is_configured": false, 00:32:37.700 "data_offset": 0, 00:32:37.700 "data_size": 65536 00:32:37.700 }, 00:32:37.700 { 00:32:37.700 "name": "BaseBdev3", 00:32:37.700 "uuid": "6a0e8948-ca88-432c-9f83-8b3698a2e7be", 00:32:37.700 "is_configured": true, 00:32:37.700 "data_offset": 0, 00:32:37.700 "data_size": 65536 00:32:37.700 } 00:32:37.700 ] 00:32:37.700 }' 00:32:37.700 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:37.700 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:37.962 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:37.962 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:37.962 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:37.962 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:32:37.962 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:37.962 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:32:37.962 13:52:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:32:37.962 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:37.962 13:52:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:37.962 [2024-11-20 13:52:45.005455] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:32:37.962 BaseBdev1 00:32:37.962 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:37.962 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:32:37.962 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:32:37.962 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:32:37.962 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:32:37.962 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:32:37.962 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:32:37.962 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:32:37.962 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:37.962 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:37.962 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:37.962 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:32:37.962 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:37.962 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:38.223 [ 00:32:38.223 { 00:32:38.223 "name": "BaseBdev1", 00:32:38.223 "aliases": [ 00:32:38.223 "b2d76513-ef62-4aac-81db-9b19a984741f" 00:32:38.223 ], 00:32:38.223 "product_name": "Malloc disk", 00:32:38.223 "block_size": 512, 00:32:38.223 "num_blocks": 65536, 00:32:38.223 "uuid": "b2d76513-ef62-4aac-81db-9b19a984741f", 00:32:38.223 "assigned_rate_limits": { 00:32:38.223 "rw_ios_per_sec": 0, 00:32:38.223 "rw_mbytes_per_sec": 0, 00:32:38.223 "r_mbytes_per_sec": 0, 00:32:38.223 "w_mbytes_per_sec": 0 00:32:38.223 }, 00:32:38.223 "claimed": true, 00:32:38.223 "claim_type": "exclusive_write", 00:32:38.223 "zoned": false, 00:32:38.223 "supported_io_types": { 00:32:38.223 "read": true, 00:32:38.223 "write": true, 00:32:38.223 "unmap": true, 00:32:38.223 "flush": true, 00:32:38.223 "reset": true, 00:32:38.223 "nvme_admin": false, 00:32:38.223 "nvme_io": false, 00:32:38.223 "nvme_io_md": false, 00:32:38.223 "write_zeroes": true, 00:32:38.223 "zcopy": true, 00:32:38.223 "get_zone_info": false, 00:32:38.223 "zone_management": false, 00:32:38.223 "zone_append": false, 00:32:38.223 "compare": false, 00:32:38.223 "compare_and_write": false, 00:32:38.223 "abort": true, 00:32:38.223 "seek_hole": false, 00:32:38.223 "seek_data": false, 00:32:38.223 "copy": true, 00:32:38.223 "nvme_iov_md": false 00:32:38.223 }, 00:32:38.223 "memory_domains": [ 00:32:38.223 { 00:32:38.223 "dma_device_id": "system", 00:32:38.223 "dma_device_type": 1 00:32:38.223 }, 00:32:38.223 { 00:32:38.223 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:32:38.223 "dma_device_type": 2 00:32:38.223 } 00:32:38.223 ], 00:32:38.223 "driver_specific": {} 00:32:38.223 } 00:32:38.223 ] 00:32:38.223 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:38.223 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:32:38.223 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:32:38.223 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:38.223 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:38.223 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:38.223 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:38.223 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:38.223 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:38.223 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:38.223 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:38.223 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:38.223 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:38.223 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:38.223 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:38.223 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:38.223 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:38.223 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:38.223 "name": "Existed_Raid", 00:32:38.223 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:38.223 "strip_size_kb": 64, 00:32:38.223 "state": "configuring", 00:32:38.223 "raid_level": "raid5f", 00:32:38.223 "superblock": false, 00:32:38.223 "num_base_bdevs": 3, 00:32:38.223 "num_base_bdevs_discovered": 2, 00:32:38.223 "num_base_bdevs_operational": 3, 00:32:38.223 "base_bdevs_list": [ 00:32:38.223 { 00:32:38.223 "name": "BaseBdev1", 00:32:38.223 "uuid": "b2d76513-ef62-4aac-81db-9b19a984741f", 00:32:38.223 "is_configured": true, 00:32:38.223 "data_offset": 0, 00:32:38.223 "data_size": 65536 00:32:38.223 }, 00:32:38.223 { 00:32:38.223 "name": null, 00:32:38.223 "uuid": "dfa6467f-25ae-44e2-877c-ff7200bf1783", 00:32:38.223 "is_configured": false, 00:32:38.223 "data_offset": 0, 00:32:38.223 "data_size": 65536 00:32:38.223 }, 00:32:38.223 { 00:32:38.223 "name": "BaseBdev3", 00:32:38.223 "uuid": "6a0e8948-ca88-432c-9f83-8b3698a2e7be", 00:32:38.223 "is_configured": true, 00:32:38.223 "data_offset": 0, 00:32:38.223 "data_size": 65536 00:32:38.223 } 00:32:38.223 ] 00:32:38.223 }' 00:32:38.223 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:38.223 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:38.481 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:38.481 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:38.481 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:38.482 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:32:38.482 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:38.482 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:32:38.482 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:32:38.482 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:38.482 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:38.482 [2024-11-20 13:52:45.397640] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:32:38.482 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:38.482 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:32:38.482 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:38.482 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:38.482 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:38.482 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:38.482 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:38.482 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:38.482 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:38.482 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:38.482 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:38.482 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:38.482 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:38.482 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:38.482 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:38.482 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:38.482 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:38.482 "name": "Existed_Raid", 00:32:38.482 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:38.482 "strip_size_kb": 64, 00:32:38.482 "state": "configuring", 00:32:38.482 "raid_level": "raid5f", 00:32:38.482 "superblock": false, 00:32:38.482 "num_base_bdevs": 3, 00:32:38.482 "num_base_bdevs_discovered": 1, 00:32:38.482 "num_base_bdevs_operational": 3, 00:32:38.482 "base_bdevs_list": [ 00:32:38.482 { 00:32:38.482 "name": "BaseBdev1", 00:32:38.482 "uuid": "b2d76513-ef62-4aac-81db-9b19a984741f", 00:32:38.482 "is_configured": true, 00:32:38.482 "data_offset": 0, 00:32:38.482 "data_size": 65536 00:32:38.482 }, 00:32:38.482 { 00:32:38.482 "name": null, 00:32:38.482 "uuid": "dfa6467f-25ae-44e2-877c-ff7200bf1783", 00:32:38.482 "is_configured": false, 00:32:38.482 "data_offset": 0, 00:32:38.482 "data_size": 65536 00:32:38.482 }, 00:32:38.482 { 00:32:38.482 "name": null, 00:32:38.482 "uuid": "6a0e8948-ca88-432c-9f83-8b3698a2e7be", 00:32:38.482 "is_configured": false, 00:32:38.482 "data_offset": 0, 00:32:38.482 "data_size": 65536 00:32:38.482 } 00:32:38.482 ] 00:32:38.482 }' 00:32:38.482 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:38.482 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:38.740 [2024-11-20 13:52:45.745730] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:38.740 "name": "Existed_Raid", 00:32:38.740 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:38.740 "strip_size_kb": 64, 00:32:38.740 "state": "configuring", 00:32:38.740 "raid_level": "raid5f", 00:32:38.740 "superblock": false, 00:32:38.740 "num_base_bdevs": 3, 00:32:38.740 "num_base_bdevs_discovered": 2, 00:32:38.740 "num_base_bdevs_operational": 3, 00:32:38.740 "base_bdevs_list": [ 00:32:38.740 { 00:32:38.740 "name": "BaseBdev1", 00:32:38.740 "uuid": "b2d76513-ef62-4aac-81db-9b19a984741f", 00:32:38.740 "is_configured": true, 00:32:38.740 "data_offset": 0, 00:32:38.740 "data_size": 65536 00:32:38.740 }, 00:32:38.740 { 00:32:38.740 "name": null, 00:32:38.740 "uuid": "dfa6467f-25ae-44e2-877c-ff7200bf1783", 00:32:38.740 "is_configured": false, 00:32:38.740 "data_offset": 0, 00:32:38.740 "data_size": 65536 00:32:38.740 }, 00:32:38.740 { 00:32:38.740 "name": "BaseBdev3", 00:32:38.740 "uuid": "6a0e8948-ca88-432c-9f83-8b3698a2e7be", 00:32:38.740 "is_configured": true, 00:32:38.740 "data_offset": 0, 00:32:38.740 "data_size": 65536 00:32:38.740 } 00:32:38.740 ] 00:32:38.740 }' 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:38.740 13:52:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:39.307 [2024-11-20 13:52:46.089805] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:39.307 "name": "Existed_Raid", 00:32:39.307 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:39.307 "strip_size_kb": 64, 00:32:39.307 "state": "configuring", 00:32:39.307 "raid_level": "raid5f", 00:32:39.307 "superblock": false, 00:32:39.307 "num_base_bdevs": 3, 00:32:39.307 "num_base_bdevs_discovered": 1, 00:32:39.307 "num_base_bdevs_operational": 3, 00:32:39.307 "base_bdevs_list": [ 00:32:39.307 { 00:32:39.307 "name": null, 00:32:39.307 "uuid": "b2d76513-ef62-4aac-81db-9b19a984741f", 00:32:39.307 "is_configured": false, 00:32:39.307 "data_offset": 0, 00:32:39.307 "data_size": 65536 00:32:39.307 }, 00:32:39.307 { 00:32:39.307 "name": null, 00:32:39.307 "uuid": "dfa6467f-25ae-44e2-877c-ff7200bf1783", 00:32:39.307 "is_configured": false, 00:32:39.307 "data_offset": 0, 00:32:39.307 "data_size": 65536 00:32:39.307 }, 00:32:39.307 { 00:32:39.307 "name": "BaseBdev3", 00:32:39.307 "uuid": "6a0e8948-ca88-432c-9f83-8b3698a2e7be", 00:32:39.307 "is_configured": true, 00:32:39.307 "data_offset": 0, 00:32:39.307 "data_size": 65536 00:32:39.307 } 00:32:39.307 ] 00:32:39.307 }' 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:39.307 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:39.566 [2024-11-20 13:52:46.516873] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:39.566 "name": "Existed_Raid", 00:32:39.566 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:39.566 "strip_size_kb": 64, 00:32:39.566 "state": "configuring", 00:32:39.566 "raid_level": "raid5f", 00:32:39.566 "superblock": false, 00:32:39.566 "num_base_bdevs": 3, 00:32:39.566 "num_base_bdevs_discovered": 2, 00:32:39.566 "num_base_bdevs_operational": 3, 00:32:39.566 "base_bdevs_list": [ 00:32:39.566 { 00:32:39.566 "name": null, 00:32:39.566 "uuid": "b2d76513-ef62-4aac-81db-9b19a984741f", 00:32:39.566 "is_configured": false, 00:32:39.566 "data_offset": 0, 00:32:39.566 "data_size": 65536 00:32:39.566 }, 00:32:39.566 { 00:32:39.566 "name": "BaseBdev2", 00:32:39.566 "uuid": "dfa6467f-25ae-44e2-877c-ff7200bf1783", 00:32:39.566 "is_configured": true, 00:32:39.566 "data_offset": 0, 00:32:39.566 "data_size": 65536 00:32:39.566 }, 00:32:39.566 { 00:32:39.566 "name": "BaseBdev3", 00:32:39.566 "uuid": "6a0e8948-ca88-432c-9f83-8b3698a2e7be", 00:32:39.566 "is_configured": true, 00:32:39.566 "data_offset": 0, 00:32:39.566 "data_size": 65536 00:32:39.566 } 00:32:39.566 ] 00:32:39.566 }' 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:39.566 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:39.824 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:39.825 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:39.825 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:39.825 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:32:39.825 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:39.825 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:32:39.825 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:39.825 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:39.825 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:39.825 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u b2d76513-ef62-4aac-81db-9b19a984741f 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:40.083 [2024-11-20 13:52:46.929479] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:32:40.083 [2024-11-20 13:52:46.929523] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:32:40.083 [2024-11-20 13:52:46.929531] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:32:40.083 [2024-11-20 13:52:46.929747] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:32:40.083 [2024-11-20 13:52:46.932700] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:32:40.083 [2024-11-20 13:52:46.932717] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:32:40.083 [2024-11-20 13:52:46.932930] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:40.083 NewBaseBdev 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:40.083 [ 00:32:40.083 { 00:32:40.083 "name": "NewBaseBdev", 00:32:40.083 "aliases": [ 00:32:40.083 "b2d76513-ef62-4aac-81db-9b19a984741f" 00:32:40.083 ], 00:32:40.083 "product_name": "Malloc disk", 00:32:40.083 "block_size": 512, 00:32:40.083 "num_blocks": 65536, 00:32:40.083 "uuid": "b2d76513-ef62-4aac-81db-9b19a984741f", 00:32:40.083 "assigned_rate_limits": { 00:32:40.083 "rw_ios_per_sec": 0, 00:32:40.083 "rw_mbytes_per_sec": 0, 00:32:40.083 "r_mbytes_per_sec": 0, 00:32:40.083 "w_mbytes_per_sec": 0 00:32:40.083 }, 00:32:40.083 "claimed": true, 00:32:40.083 "claim_type": "exclusive_write", 00:32:40.083 "zoned": false, 00:32:40.083 "supported_io_types": { 00:32:40.083 "read": true, 00:32:40.083 "write": true, 00:32:40.083 "unmap": true, 00:32:40.083 "flush": true, 00:32:40.083 "reset": true, 00:32:40.083 "nvme_admin": false, 00:32:40.083 "nvme_io": false, 00:32:40.083 "nvme_io_md": false, 00:32:40.083 "write_zeroes": true, 00:32:40.083 "zcopy": true, 00:32:40.083 "get_zone_info": false, 00:32:40.083 "zone_management": false, 00:32:40.083 "zone_append": false, 00:32:40.083 "compare": false, 00:32:40.083 "compare_and_write": false, 00:32:40.083 "abort": true, 00:32:40.083 "seek_hole": false, 00:32:40.083 "seek_data": false, 00:32:40.083 "copy": true, 00:32:40.083 "nvme_iov_md": false 00:32:40.083 }, 00:32:40.083 "memory_domains": [ 00:32:40.083 { 00:32:40.083 "dma_device_id": "system", 00:32:40.083 "dma_device_type": 1 00:32:40.083 }, 00:32:40.083 { 00:32:40.083 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:32:40.083 "dma_device_type": 2 00:32:40.083 } 00:32:40.083 ], 00:32:40.083 "driver_specific": {} 00:32:40.083 } 00:32:40.083 ] 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:40.083 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:40.083 "name": "Existed_Raid", 00:32:40.083 "uuid": "83f22e2a-abcd-4f31-8366-5668fa38bad4", 00:32:40.083 "strip_size_kb": 64, 00:32:40.083 "state": "online", 00:32:40.083 "raid_level": "raid5f", 00:32:40.083 "superblock": false, 00:32:40.083 "num_base_bdevs": 3, 00:32:40.083 "num_base_bdevs_discovered": 3, 00:32:40.083 "num_base_bdevs_operational": 3, 00:32:40.083 "base_bdevs_list": [ 00:32:40.083 { 00:32:40.083 "name": "NewBaseBdev", 00:32:40.083 "uuid": "b2d76513-ef62-4aac-81db-9b19a984741f", 00:32:40.083 "is_configured": true, 00:32:40.084 "data_offset": 0, 00:32:40.084 "data_size": 65536 00:32:40.084 }, 00:32:40.084 { 00:32:40.084 "name": "BaseBdev2", 00:32:40.084 "uuid": "dfa6467f-25ae-44e2-877c-ff7200bf1783", 00:32:40.084 "is_configured": true, 00:32:40.084 "data_offset": 0, 00:32:40.084 "data_size": 65536 00:32:40.084 }, 00:32:40.084 { 00:32:40.084 "name": "BaseBdev3", 00:32:40.084 "uuid": "6a0e8948-ca88-432c-9f83-8b3698a2e7be", 00:32:40.084 "is_configured": true, 00:32:40.084 "data_offset": 0, 00:32:40.084 "data_size": 65536 00:32:40.084 } 00:32:40.084 ] 00:32:40.084 }' 00:32:40.084 13:52:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:40.084 13:52:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:40.343 [2024-11-20 13:52:47.264767] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:32:40.343 "name": "Existed_Raid", 00:32:40.343 "aliases": [ 00:32:40.343 "83f22e2a-abcd-4f31-8366-5668fa38bad4" 00:32:40.343 ], 00:32:40.343 "product_name": "Raid Volume", 00:32:40.343 "block_size": 512, 00:32:40.343 "num_blocks": 131072, 00:32:40.343 "uuid": "83f22e2a-abcd-4f31-8366-5668fa38bad4", 00:32:40.343 "assigned_rate_limits": { 00:32:40.343 "rw_ios_per_sec": 0, 00:32:40.343 "rw_mbytes_per_sec": 0, 00:32:40.343 "r_mbytes_per_sec": 0, 00:32:40.343 "w_mbytes_per_sec": 0 00:32:40.343 }, 00:32:40.343 "claimed": false, 00:32:40.343 "zoned": false, 00:32:40.343 "supported_io_types": { 00:32:40.343 "read": true, 00:32:40.343 "write": true, 00:32:40.343 "unmap": false, 00:32:40.343 "flush": false, 00:32:40.343 "reset": true, 00:32:40.343 "nvme_admin": false, 00:32:40.343 "nvme_io": false, 00:32:40.343 "nvme_io_md": false, 00:32:40.343 "write_zeroes": true, 00:32:40.343 "zcopy": false, 00:32:40.343 "get_zone_info": false, 00:32:40.343 "zone_management": false, 00:32:40.343 "zone_append": false, 00:32:40.343 "compare": false, 00:32:40.343 "compare_and_write": false, 00:32:40.343 "abort": false, 00:32:40.343 "seek_hole": false, 00:32:40.343 "seek_data": false, 00:32:40.343 "copy": false, 00:32:40.343 "nvme_iov_md": false 00:32:40.343 }, 00:32:40.343 "driver_specific": { 00:32:40.343 "raid": { 00:32:40.343 "uuid": "83f22e2a-abcd-4f31-8366-5668fa38bad4", 00:32:40.343 "strip_size_kb": 64, 00:32:40.343 "state": "online", 00:32:40.343 "raid_level": "raid5f", 00:32:40.343 "superblock": false, 00:32:40.343 "num_base_bdevs": 3, 00:32:40.343 "num_base_bdevs_discovered": 3, 00:32:40.343 "num_base_bdevs_operational": 3, 00:32:40.343 "base_bdevs_list": [ 00:32:40.343 { 00:32:40.343 "name": "NewBaseBdev", 00:32:40.343 "uuid": "b2d76513-ef62-4aac-81db-9b19a984741f", 00:32:40.343 "is_configured": true, 00:32:40.343 "data_offset": 0, 00:32:40.343 "data_size": 65536 00:32:40.343 }, 00:32:40.343 { 00:32:40.343 "name": "BaseBdev2", 00:32:40.343 "uuid": "dfa6467f-25ae-44e2-877c-ff7200bf1783", 00:32:40.343 "is_configured": true, 00:32:40.343 "data_offset": 0, 00:32:40.343 "data_size": 65536 00:32:40.343 }, 00:32:40.343 { 00:32:40.343 "name": "BaseBdev3", 00:32:40.343 "uuid": "6a0e8948-ca88-432c-9f83-8b3698a2e7be", 00:32:40.343 "is_configured": true, 00:32:40.343 "data_offset": 0, 00:32:40.343 "data_size": 65536 00:32:40.343 } 00:32:40.343 ] 00:32:40.343 } 00:32:40.343 } 00:32:40.343 }' 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:32:40.343 BaseBdev2 00:32:40.343 BaseBdev3' 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:40.343 13:52:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:40.601 13:52:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:40.601 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:32:40.601 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:32:40.601 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:32:40.601 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:40.601 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:32:40.601 13:52:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:40.601 13:52:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:40.601 13:52:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:40.601 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:32:40.601 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:32:40.601 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:32:40.601 13:52:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:40.601 13:52:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:40.601 [2024-11-20 13:52:47.468615] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:32:40.602 [2024-11-20 13:52:47.468738] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:32:40.602 [2024-11-20 13:52:47.468830] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:32:40.602 [2024-11-20 13:52:47.469092] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:32:40.602 [2024-11-20 13:52:47.469105] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:32:40.602 13:52:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:40.602 13:52:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 77637 00:32:40.602 13:52:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 77637 ']' 00:32:40.602 13:52:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@958 -- # kill -0 77637 00:32:40.602 13:52:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # uname 00:32:40.602 13:52:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:32:40.602 13:52:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 77637 00:32:40.602 13:52:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:32:40.602 13:52:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:32:40.602 13:52:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 77637' 00:32:40.602 killing process with pid 77637 00:32:40.602 13:52:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@973 -- # kill 77637 00:32:40.602 [2024-11-20 13:52:47.498171] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:32:40.602 13:52:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@978 -- # wait 77637 00:32:40.602 [2024-11-20 13:52:47.656229] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:32:41.538 13:52:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:32:41.538 00:32:41.538 real 0m7.722s 00:32:41.538 user 0m12.232s 00:32:41.538 sys 0m1.388s 00:32:41.538 13:52:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:32:41.538 13:52:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:32:41.538 ************************************ 00:32:41.538 END TEST raid5f_state_function_test 00:32:41.538 ************************************ 00:32:41.538 13:52:48 bdev_raid -- bdev/bdev_raid.sh@987 -- # run_test raid5f_state_function_test_sb raid_state_function_test raid5f 3 true 00:32:41.538 13:52:48 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:32:41.538 13:52:48 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:32:41.538 13:52:48 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:32:41.538 ************************************ 00:32:41.538 START TEST raid5f_state_function_test_sb 00:32:41.538 ************************************ 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 3 true 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:32:41.539 Process raid pid: 78232 00:32:41.539 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=78232 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 78232' 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 78232 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 78232 ']' 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:41.539 13:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:32:41.539 [2024-11-20 13:52:48.405190] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:32:41.539 [2024-11-20 13:52:48.405872] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:32:41.539 [2024-11-20 13:52:48.565712] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:41.797 [2024-11-20 13:52:48.684631] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:32:41.797 [2024-11-20 13:52:48.835723] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:32:41.797 [2024-11-20 13:52:48.835775] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:32:42.375 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:32:42.375 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:32:42.375 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:32:42.375 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:42.375 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:42.375 [2024-11-20 13:52:49.233546] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:32:42.375 [2024-11-20 13:52:49.233611] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:32:42.375 [2024-11-20 13:52:49.233623] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:32:42.375 [2024-11-20 13:52:49.233633] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:32:42.375 [2024-11-20 13:52:49.233639] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:32:42.375 [2024-11-20 13:52:49.233648] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:32:42.375 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:42.375 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:32:42.375 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:42.375 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:42.375 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:42.375 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:42.375 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:42.375 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:42.375 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:42.375 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:42.375 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:42.375 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:42.375 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:42.375 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:42.375 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:42.375 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:42.375 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:42.375 "name": "Existed_Raid", 00:32:42.375 "uuid": "3aa62224-7182-44c6-929b-db35d43f755e", 00:32:42.375 "strip_size_kb": 64, 00:32:42.375 "state": "configuring", 00:32:42.375 "raid_level": "raid5f", 00:32:42.375 "superblock": true, 00:32:42.375 "num_base_bdevs": 3, 00:32:42.375 "num_base_bdevs_discovered": 0, 00:32:42.375 "num_base_bdevs_operational": 3, 00:32:42.375 "base_bdevs_list": [ 00:32:42.375 { 00:32:42.375 "name": "BaseBdev1", 00:32:42.375 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:42.375 "is_configured": false, 00:32:42.375 "data_offset": 0, 00:32:42.375 "data_size": 0 00:32:42.375 }, 00:32:42.375 { 00:32:42.375 "name": "BaseBdev2", 00:32:42.375 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:42.375 "is_configured": false, 00:32:42.375 "data_offset": 0, 00:32:42.375 "data_size": 0 00:32:42.375 }, 00:32:42.375 { 00:32:42.375 "name": "BaseBdev3", 00:32:42.375 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:42.375 "is_configured": false, 00:32:42.375 "data_offset": 0, 00:32:42.375 "data_size": 0 00:32:42.375 } 00:32:42.375 ] 00:32:42.375 }' 00:32:42.375 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:42.375 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:42.635 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:32:42.635 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:42.635 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:42.635 [2024-11-20 13:52:49.533555] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:32:42.635 [2024-11-20 13:52:49.533602] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:42.636 [2024-11-20 13:52:49.541555] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:32:42.636 [2024-11-20 13:52:49.541603] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:32:42.636 [2024-11-20 13:52:49.541612] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:32:42.636 [2024-11-20 13:52:49.541622] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:32:42.636 [2024-11-20 13:52:49.541629] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:32:42.636 [2024-11-20 13:52:49.541638] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:42.636 [2024-11-20 13:52:49.576304] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:32:42.636 BaseBdev1 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:42.636 [ 00:32:42.636 { 00:32:42.636 "name": "BaseBdev1", 00:32:42.636 "aliases": [ 00:32:42.636 "39d55e21-8d65-4ef1-a840-d5e7ce9a124b" 00:32:42.636 ], 00:32:42.636 "product_name": "Malloc disk", 00:32:42.636 "block_size": 512, 00:32:42.636 "num_blocks": 65536, 00:32:42.636 "uuid": "39d55e21-8d65-4ef1-a840-d5e7ce9a124b", 00:32:42.636 "assigned_rate_limits": { 00:32:42.636 "rw_ios_per_sec": 0, 00:32:42.636 "rw_mbytes_per_sec": 0, 00:32:42.636 "r_mbytes_per_sec": 0, 00:32:42.636 "w_mbytes_per_sec": 0 00:32:42.636 }, 00:32:42.636 "claimed": true, 00:32:42.636 "claim_type": "exclusive_write", 00:32:42.636 "zoned": false, 00:32:42.636 "supported_io_types": { 00:32:42.636 "read": true, 00:32:42.636 "write": true, 00:32:42.636 "unmap": true, 00:32:42.636 "flush": true, 00:32:42.636 "reset": true, 00:32:42.636 "nvme_admin": false, 00:32:42.636 "nvme_io": false, 00:32:42.636 "nvme_io_md": false, 00:32:42.636 "write_zeroes": true, 00:32:42.636 "zcopy": true, 00:32:42.636 "get_zone_info": false, 00:32:42.636 "zone_management": false, 00:32:42.636 "zone_append": false, 00:32:42.636 "compare": false, 00:32:42.636 "compare_and_write": false, 00:32:42.636 "abort": true, 00:32:42.636 "seek_hole": false, 00:32:42.636 "seek_data": false, 00:32:42.636 "copy": true, 00:32:42.636 "nvme_iov_md": false 00:32:42.636 }, 00:32:42.636 "memory_domains": [ 00:32:42.636 { 00:32:42.636 "dma_device_id": "system", 00:32:42.636 "dma_device_type": 1 00:32:42.636 }, 00:32:42.636 { 00:32:42.636 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:32:42.636 "dma_device_type": 2 00:32:42.636 } 00:32:42.636 ], 00:32:42.636 "driver_specific": {} 00:32:42.636 } 00:32:42.636 ] 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:42.636 "name": "Existed_Raid", 00:32:42.636 "uuid": "688f21fe-11ab-47d7-a2e2-a284cab4f56f", 00:32:42.636 "strip_size_kb": 64, 00:32:42.636 "state": "configuring", 00:32:42.636 "raid_level": "raid5f", 00:32:42.636 "superblock": true, 00:32:42.636 "num_base_bdevs": 3, 00:32:42.636 "num_base_bdevs_discovered": 1, 00:32:42.636 "num_base_bdevs_operational": 3, 00:32:42.636 "base_bdevs_list": [ 00:32:42.636 { 00:32:42.636 "name": "BaseBdev1", 00:32:42.636 "uuid": "39d55e21-8d65-4ef1-a840-d5e7ce9a124b", 00:32:42.636 "is_configured": true, 00:32:42.636 "data_offset": 2048, 00:32:42.636 "data_size": 63488 00:32:42.636 }, 00:32:42.636 { 00:32:42.636 "name": "BaseBdev2", 00:32:42.636 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:42.636 "is_configured": false, 00:32:42.636 "data_offset": 0, 00:32:42.636 "data_size": 0 00:32:42.636 }, 00:32:42.636 { 00:32:42.636 "name": "BaseBdev3", 00:32:42.636 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:42.636 "is_configured": false, 00:32:42.636 "data_offset": 0, 00:32:42.636 "data_size": 0 00:32:42.636 } 00:32:42.636 ] 00:32:42.636 }' 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:42.636 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:42.962 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:32:42.962 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:42.962 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:42.962 [2024-11-20 13:52:49.920450] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:32:42.962 [2024-11-20 13:52:49.920514] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:32:42.962 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:42.962 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:32:42.962 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:42.962 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:42.962 [2024-11-20 13:52:49.928500] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:32:42.962 [2024-11-20 13:52:49.930482] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:32:42.962 [2024-11-20 13:52:49.930522] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:32:42.962 [2024-11-20 13:52:49.930532] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:32:42.962 [2024-11-20 13:52:49.930541] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:32:42.962 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:42.962 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:32:42.962 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:32:42.962 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:32:42.962 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:42.962 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:42.962 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:42.962 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:42.962 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:42.962 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:42.962 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:42.962 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:42.962 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:42.962 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:42.962 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:42.962 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:42.962 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:42.962 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:42.962 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:42.962 "name": "Existed_Raid", 00:32:42.962 "uuid": "91a384d2-8071-469f-9886-cf59b9bd1f08", 00:32:42.962 "strip_size_kb": 64, 00:32:42.962 "state": "configuring", 00:32:42.962 "raid_level": "raid5f", 00:32:42.962 "superblock": true, 00:32:42.962 "num_base_bdevs": 3, 00:32:42.962 "num_base_bdevs_discovered": 1, 00:32:42.962 "num_base_bdevs_operational": 3, 00:32:42.962 "base_bdevs_list": [ 00:32:42.962 { 00:32:42.962 "name": "BaseBdev1", 00:32:42.962 "uuid": "39d55e21-8d65-4ef1-a840-d5e7ce9a124b", 00:32:42.962 "is_configured": true, 00:32:42.962 "data_offset": 2048, 00:32:42.962 "data_size": 63488 00:32:42.962 }, 00:32:42.962 { 00:32:42.962 "name": "BaseBdev2", 00:32:42.962 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:42.962 "is_configured": false, 00:32:42.962 "data_offset": 0, 00:32:42.962 "data_size": 0 00:32:42.962 }, 00:32:42.962 { 00:32:42.962 "name": "BaseBdev3", 00:32:42.962 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:42.962 "is_configured": false, 00:32:42.962 "data_offset": 0, 00:32:42.962 "data_size": 0 00:32:42.962 } 00:32:42.962 ] 00:32:42.962 }' 00:32:42.963 13:52:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:42.963 13:52:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:43.231 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:32:43.231 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:43.231 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:43.231 [2024-11-20 13:52:50.265196] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:32:43.231 BaseBdev2 00:32:43.231 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:43.231 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:32:43.231 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:32:43.231 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:32:43.231 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:32:43.231 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:32:43.231 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:32:43.231 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:32:43.231 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:43.231 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:43.231 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:43.231 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:32:43.231 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:43.231 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:43.489 [ 00:32:43.489 { 00:32:43.489 "name": "BaseBdev2", 00:32:43.489 "aliases": [ 00:32:43.489 "ae1935f1-ff26-40cb-b00b-72cc526d6c4f" 00:32:43.489 ], 00:32:43.489 "product_name": "Malloc disk", 00:32:43.489 "block_size": 512, 00:32:43.489 "num_blocks": 65536, 00:32:43.489 "uuid": "ae1935f1-ff26-40cb-b00b-72cc526d6c4f", 00:32:43.489 "assigned_rate_limits": { 00:32:43.489 "rw_ios_per_sec": 0, 00:32:43.489 "rw_mbytes_per_sec": 0, 00:32:43.489 "r_mbytes_per_sec": 0, 00:32:43.489 "w_mbytes_per_sec": 0 00:32:43.489 }, 00:32:43.489 "claimed": true, 00:32:43.489 "claim_type": "exclusive_write", 00:32:43.489 "zoned": false, 00:32:43.489 "supported_io_types": { 00:32:43.489 "read": true, 00:32:43.489 "write": true, 00:32:43.489 "unmap": true, 00:32:43.489 "flush": true, 00:32:43.489 "reset": true, 00:32:43.489 "nvme_admin": false, 00:32:43.489 "nvme_io": false, 00:32:43.489 "nvme_io_md": false, 00:32:43.489 "write_zeroes": true, 00:32:43.489 "zcopy": true, 00:32:43.489 "get_zone_info": false, 00:32:43.489 "zone_management": false, 00:32:43.489 "zone_append": false, 00:32:43.489 "compare": false, 00:32:43.489 "compare_and_write": false, 00:32:43.489 "abort": true, 00:32:43.489 "seek_hole": false, 00:32:43.489 "seek_data": false, 00:32:43.489 "copy": true, 00:32:43.489 "nvme_iov_md": false 00:32:43.489 }, 00:32:43.489 "memory_domains": [ 00:32:43.489 { 00:32:43.489 "dma_device_id": "system", 00:32:43.489 "dma_device_type": 1 00:32:43.489 }, 00:32:43.489 { 00:32:43.489 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:32:43.489 "dma_device_type": 2 00:32:43.489 } 00:32:43.489 ], 00:32:43.489 "driver_specific": {} 00:32:43.489 } 00:32:43.489 ] 00:32:43.489 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:43.489 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:32:43.489 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:32:43.489 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:32:43.489 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:32:43.489 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:43.489 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:43.489 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:43.489 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:43.489 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:43.489 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:43.489 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:43.489 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:43.490 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:43.490 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:43.490 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:43.490 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:43.490 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:43.490 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:43.490 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:43.490 "name": "Existed_Raid", 00:32:43.490 "uuid": "91a384d2-8071-469f-9886-cf59b9bd1f08", 00:32:43.490 "strip_size_kb": 64, 00:32:43.490 "state": "configuring", 00:32:43.490 "raid_level": "raid5f", 00:32:43.490 "superblock": true, 00:32:43.490 "num_base_bdevs": 3, 00:32:43.490 "num_base_bdevs_discovered": 2, 00:32:43.490 "num_base_bdevs_operational": 3, 00:32:43.490 "base_bdevs_list": [ 00:32:43.490 { 00:32:43.490 "name": "BaseBdev1", 00:32:43.490 "uuid": "39d55e21-8d65-4ef1-a840-d5e7ce9a124b", 00:32:43.490 "is_configured": true, 00:32:43.490 "data_offset": 2048, 00:32:43.490 "data_size": 63488 00:32:43.490 }, 00:32:43.490 { 00:32:43.490 "name": "BaseBdev2", 00:32:43.490 "uuid": "ae1935f1-ff26-40cb-b00b-72cc526d6c4f", 00:32:43.490 "is_configured": true, 00:32:43.490 "data_offset": 2048, 00:32:43.490 "data_size": 63488 00:32:43.490 }, 00:32:43.490 { 00:32:43.490 "name": "BaseBdev3", 00:32:43.490 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:43.490 "is_configured": false, 00:32:43.490 "data_offset": 0, 00:32:43.490 "data_size": 0 00:32:43.490 } 00:32:43.490 ] 00:32:43.490 }' 00:32:43.490 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:43.490 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:43.748 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:32:43.748 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:43.748 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:43.748 [2024-11-20 13:52:50.624699] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:32:43.748 [2024-11-20 13:52:50.624953] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:32:43.748 [2024-11-20 13:52:50.624972] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:32:43.748 BaseBdev3 00:32:43.748 [2024-11-20 13:52:50.625271] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:32:43.748 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:43.748 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:32:43.748 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:32:43.748 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:32:43.748 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:32:43.748 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:32:43.748 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:32:43.748 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:32:43.748 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:43.748 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:43.748 [2024-11-20 13:52:50.629116] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:32:43.748 [2024-11-20 13:52:50.629136] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:32:43.748 [2024-11-20 13:52:50.629389] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:43.748 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:43.748 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:32:43.748 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:43.748 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:43.748 [ 00:32:43.748 { 00:32:43.748 "name": "BaseBdev3", 00:32:43.748 "aliases": [ 00:32:43.748 "9e5244f3-fd85-433f-a00d-baca001cc4c9" 00:32:43.748 ], 00:32:43.748 "product_name": "Malloc disk", 00:32:43.748 "block_size": 512, 00:32:43.748 "num_blocks": 65536, 00:32:43.748 "uuid": "9e5244f3-fd85-433f-a00d-baca001cc4c9", 00:32:43.748 "assigned_rate_limits": { 00:32:43.748 "rw_ios_per_sec": 0, 00:32:43.749 "rw_mbytes_per_sec": 0, 00:32:43.749 "r_mbytes_per_sec": 0, 00:32:43.749 "w_mbytes_per_sec": 0 00:32:43.749 }, 00:32:43.749 "claimed": true, 00:32:43.749 "claim_type": "exclusive_write", 00:32:43.749 "zoned": false, 00:32:43.749 "supported_io_types": { 00:32:43.749 "read": true, 00:32:43.749 "write": true, 00:32:43.749 "unmap": true, 00:32:43.749 "flush": true, 00:32:43.749 "reset": true, 00:32:43.749 "nvme_admin": false, 00:32:43.749 "nvme_io": false, 00:32:43.749 "nvme_io_md": false, 00:32:43.749 "write_zeroes": true, 00:32:43.749 "zcopy": true, 00:32:43.749 "get_zone_info": false, 00:32:43.749 "zone_management": false, 00:32:43.749 "zone_append": false, 00:32:43.749 "compare": false, 00:32:43.749 "compare_and_write": false, 00:32:43.749 "abort": true, 00:32:43.749 "seek_hole": false, 00:32:43.749 "seek_data": false, 00:32:43.749 "copy": true, 00:32:43.749 "nvme_iov_md": false 00:32:43.749 }, 00:32:43.749 "memory_domains": [ 00:32:43.749 { 00:32:43.749 "dma_device_id": "system", 00:32:43.749 "dma_device_type": 1 00:32:43.749 }, 00:32:43.749 { 00:32:43.749 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:32:43.749 "dma_device_type": 2 00:32:43.749 } 00:32:43.749 ], 00:32:43.749 "driver_specific": {} 00:32:43.749 } 00:32:43.749 ] 00:32:43.749 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:43.749 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:32:43.749 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:32:43.749 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:32:43.749 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:32:43.749 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:43.749 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:43.749 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:43.749 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:43.749 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:43.749 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:43.749 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:43.749 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:43.749 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:43.749 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:43.749 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:43.749 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:43.749 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:43.749 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:43.749 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:43.749 "name": "Existed_Raid", 00:32:43.749 "uuid": "91a384d2-8071-469f-9886-cf59b9bd1f08", 00:32:43.749 "strip_size_kb": 64, 00:32:43.749 "state": "online", 00:32:43.749 "raid_level": "raid5f", 00:32:43.749 "superblock": true, 00:32:43.749 "num_base_bdevs": 3, 00:32:43.749 "num_base_bdevs_discovered": 3, 00:32:43.749 "num_base_bdevs_operational": 3, 00:32:43.749 "base_bdevs_list": [ 00:32:43.749 { 00:32:43.749 "name": "BaseBdev1", 00:32:43.749 "uuid": "39d55e21-8d65-4ef1-a840-d5e7ce9a124b", 00:32:43.749 "is_configured": true, 00:32:43.749 "data_offset": 2048, 00:32:43.749 "data_size": 63488 00:32:43.749 }, 00:32:43.749 { 00:32:43.749 "name": "BaseBdev2", 00:32:43.749 "uuid": "ae1935f1-ff26-40cb-b00b-72cc526d6c4f", 00:32:43.749 "is_configured": true, 00:32:43.749 "data_offset": 2048, 00:32:43.749 "data_size": 63488 00:32:43.749 }, 00:32:43.749 { 00:32:43.749 "name": "BaseBdev3", 00:32:43.749 "uuid": "9e5244f3-fd85-433f-a00d-baca001cc4c9", 00:32:43.749 "is_configured": true, 00:32:43.749 "data_offset": 2048, 00:32:43.749 "data_size": 63488 00:32:43.749 } 00:32:43.749 ] 00:32:43.749 }' 00:32:43.749 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:43.749 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:44.008 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:32:44.008 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:32:44.008 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:32:44.008 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:32:44.008 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:32:44.008 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:32:44.008 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:32:44.008 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:32:44.008 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:44.008 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:44.008 [2024-11-20 13:52:50.958141] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:32:44.008 13:52:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:44.008 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:32:44.008 "name": "Existed_Raid", 00:32:44.008 "aliases": [ 00:32:44.008 "91a384d2-8071-469f-9886-cf59b9bd1f08" 00:32:44.008 ], 00:32:44.008 "product_name": "Raid Volume", 00:32:44.008 "block_size": 512, 00:32:44.008 "num_blocks": 126976, 00:32:44.008 "uuid": "91a384d2-8071-469f-9886-cf59b9bd1f08", 00:32:44.008 "assigned_rate_limits": { 00:32:44.008 "rw_ios_per_sec": 0, 00:32:44.008 "rw_mbytes_per_sec": 0, 00:32:44.008 "r_mbytes_per_sec": 0, 00:32:44.008 "w_mbytes_per_sec": 0 00:32:44.008 }, 00:32:44.008 "claimed": false, 00:32:44.008 "zoned": false, 00:32:44.008 "supported_io_types": { 00:32:44.008 "read": true, 00:32:44.008 "write": true, 00:32:44.008 "unmap": false, 00:32:44.008 "flush": false, 00:32:44.008 "reset": true, 00:32:44.008 "nvme_admin": false, 00:32:44.008 "nvme_io": false, 00:32:44.008 "nvme_io_md": false, 00:32:44.008 "write_zeroes": true, 00:32:44.008 "zcopy": false, 00:32:44.008 "get_zone_info": false, 00:32:44.008 "zone_management": false, 00:32:44.008 "zone_append": false, 00:32:44.008 "compare": false, 00:32:44.008 "compare_and_write": false, 00:32:44.008 "abort": false, 00:32:44.008 "seek_hole": false, 00:32:44.008 "seek_data": false, 00:32:44.008 "copy": false, 00:32:44.008 "nvme_iov_md": false 00:32:44.008 }, 00:32:44.008 "driver_specific": { 00:32:44.008 "raid": { 00:32:44.008 "uuid": "91a384d2-8071-469f-9886-cf59b9bd1f08", 00:32:44.008 "strip_size_kb": 64, 00:32:44.008 "state": "online", 00:32:44.008 "raid_level": "raid5f", 00:32:44.008 "superblock": true, 00:32:44.008 "num_base_bdevs": 3, 00:32:44.008 "num_base_bdevs_discovered": 3, 00:32:44.008 "num_base_bdevs_operational": 3, 00:32:44.008 "base_bdevs_list": [ 00:32:44.008 { 00:32:44.008 "name": "BaseBdev1", 00:32:44.008 "uuid": "39d55e21-8d65-4ef1-a840-d5e7ce9a124b", 00:32:44.008 "is_configured": true, 00:32:44.008 "data_offset": 2048, 00:32:44.008 "data_size": 63488 00:32:44.008 }, 00:32:44.008 { 00:32:44.008 "name": "BaseBdev2", 00:32:44.008 "uuid": "ae1935f1-ff26-40cb-b00b-72cc526d6c4f", 00:32:44.008 "is_configured": true, 00:32:44.008 "data_offset": 2048, 00:32:44.008 "data_size": 63488 00:32:44.008 }, 00:32:44.008 { 00:32:44.008 "name": "BaseBdev3", 00:32:44.008 "uuid": "9e5244f3-fd85-433f-a00d-baca001cc4c9", 00:32:44.008 "is_configured": true, 00:32:44.008 "data_offset": 2048, 00:32:44.008 "data_size": 63488 00:32:44.008 } 00:32:44.008 ] 00:32:44.008 } 00:32:44.008 } 00:32:44.008 }' 00:32:44.008 13:52:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:32:44.008 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:32:44.008 BaseBdev2 00:32:44.008 BaseBdev3' 00:32:44.008 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:44.008 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:32:44.008 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:32:44.008 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:32:44.008 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:44.008 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:44.008 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:44.008 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:44.267 [2024-11-20 13:52:51.157961] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 2 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:44.267 "name": "Existed_Raid", 00:32:44.267 "uuid": "91a384d2-8071-469f-9886-cf59b9bd1f08", 00:32:44.267 "strip_size_kb": 64, 00:32:44.267 "state": "online", 00:32:44.267 "raid_level": "raid5f", 00:32:44.267 "superblock": true, 00:32:44.267 "num_base_bdevs": 3, 00:32:44.267 "num_base_bdevs_discovered": 2, 00:32:44.267 "num_base_bdevs_operational": 2, 00:32:44.267 "base_bdevs_list": [ 00:32:44.267 { 00:32:44.267 "name": null, 00:32:44.267 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:44.267 "is_configured": false, 00:32:44.267 "data_offset": 0, 00:32:44.267 "data_size": 63488 00:32:44.267 }, 00:32:44.267 { 00:32:44.267 "name": "BaseBdev2", 00:32:44.267 "uuid": "ae1935f1-ff26-40cb-b00b-72cc526d6c4f", 00:32:44.267 "is_configured": true, 00:32:44.267 "data_offset": 2048, 00:32:44.267 "data_size": 63488 00:32:44.267 }, 00:32:44.267 { 00:32:44.267 "name": "BaseBdev3", 00:32:44.267 "uuid": "9e5244f3-fd85-433f-a00d-baca001cc4c9", 00:32:44.267 "is_configured": true, 00:32:44.267 "data_offset": 2048, 00:32:44.267 "data_size": 63488 00:32:44.267 } 00:32:44.267 ] 00:32:44.267 }' 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:44.267 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:44.527 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:32:44.527 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:32:44.527 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:32:44.527 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:44.527 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:44.527 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:44.527 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:44.527 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:32:44.527 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:32:44.527 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:32:44.527 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:44.527 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:44.527 [2024-11-20 13:52:51.580007] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:32:44.527 [2024-11-20 13:52:51.580165] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:32:44.787 [2024-11-20 13:52:51.644179] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:32:44.787 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:44.788 [2024-11-20 13:52:51.684241] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:32:44.788 [2024-11-20 13:52:51.684294] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:44.788 BaseBdev2 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:44.788 [ 00:32:44.788 { 00:32:44.788 "name": "BaseBdev2", 00:32:44.788 "aliases": [ 00:32:44.788 "1a411ecd-5151-427a-b2e7-ce9b4d0c40bb" 00:32:44.788 ], 00:32:44.788 "product_name": "Malloc disk", 00:32:44.788 "block_size": 512, 00:32:44.788 "num_blocks": 65536, 00:32:44.788 "uuid": "1a411ecd-5151-427a-b2e7-ce9b4d0c40bb", 00:32:44.788 "assigned_rate_limits": { 00:32:44.788 "rw_ios_per_sec": 0, 00:32:44.788 "rw_mbytes_per_sec": 0, 00:32:44.788 "r_mbytes_per_sec": 0, 00:32:44.788 "w_mbytes_per_sec": 0 00:32:44.788 }, 00:32:44.788 "claimed": false, 00:32:44.788 "zoned": false, 00:32:44.788 "supported_io_types": { 00:32:44.788 "read": true, 00:32:44.788 "write": true, 00:32:44.788 "unmap": true, 00:32:44.788 "flush": true, 00:32:44.788 "reset": true, 00:32:44.788 "nvme_admin": false, 00:32:44.788 "nvme_io": false, 00:32:44.788 "nvme_io_md": false, 00:32:44.788 "write_zeroes": true, 00:32:44.788 "zcopy": true, 00:32:44.788 "get_zone_info": false, 00:32:44.788 "zone_management": false, 00:32:44.788 "zone_append": false, 00:32:44.788 "compare": false, 00:32:44.788 "compare_and_write": false, 00:32:44.788 "abort": true, 00:32:44.788 "seek_hole": false, 00:32:44.788 "seek_data": false, 00:32:44.788 "copy": true, 00:32:44.788 "nvme_iov_md": false 00:32:44.788 }, 00:32:44.788 "memory_domains": [ 00:32:44.788 { 00:32:44.788 "dma_device_id": "system", 00:32:44.788 "dma_device_type": 1 00:32:44.788 }, 00:32:44.788 { 00:32:44.788 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:32:44.788 "dma_device_type": 2 00:32:44.788 } 00:32:44.788 ], 00:32:44.788 "driver_specific": {} 00:32:44.788 } 00:32:44.788 ] 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:44.788 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:45.047 BaseBdev3 00:32:45.047 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:45.047 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:45.048 [ 00:32:45.048 { 00:32:45.048 "name": "BaseBdev3", 00:32:45.048 "aliases": [ 00:32:45.048 "cfc71633-cabf-47aa-92b3-71f72250cf01" 00:32:45.048 ], 00:32:45.048 "product_name": "Malloc disk", 00:32:45.048 "block_size": 512, 00:32:45.048 "num_blocks": 65536, 00:32:45.048 "uuid": "cfc71633-cabf-47aa-92b3-71f72250cf01", 00:32:45.048 "assigned_rate_limits": { 00:32:45.048 "rw_ios_per_sec": 0, 00:32:45.048 "rw_mbytes_per_sec": 0, 00:32:45.048 "r_mbytes_per_sec": 0, 00:32:45.048 "w_mbytes_per_sec": 0 00:32:45.048 }, 00:32:45.048 "claimed": false, 00:32:45.048 "zoned": false, 00:32:45.048 "supported_io_types": { 00:32:45.048 "read": true, 00:32:45.048 "write": true, 00:32:45.048 "unmap": true, 00:32:45.048 "flush": true, 00:32:45.048 "reset": true, 00:32:45.048 "nvme_admin": false, 00:32:45.048 "nvme_io": false, 00:32:45.048 "nvme_io_md": false, 00:32:45.048 "write_zeroes": true, 00:32:45.048 "zcopy": true, 00:32:45.048 "get_zone_info": false, 00:32:45.048 "zone_management": false, 00:32:45.048 "zone_append": false, 00:32:45.048 "compare": false, 00:32:45.048 "compare_and_write": false, 00:32:45.048 "abort": true, 00:32:45.048 "seek_hole": false, 00:32:45.048 "seek_data": false, 00:32:45.048 "copy": true, 00:32:45.048 "nvme_iov_md": false 00:32:45.048 }, 00:32:45.048 "memory_domains": [ 00:32:45.048 { 00:32:45.048 "dma_device_id": "system", 00:32:45.048 "dma_device_type": 1 00:32:45.048 }, 00:32:45.048 { 00:32:45.048 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:32:45.048 "dma_device_type": 2 00:32:45.048 } 00:32:45.048 ], 00:32:45.048 "driver_specific": {} 00:32:45.048 } 00:32:45.048 ] 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:45.048 [2024-11-20 13:52:51.905766] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:32:45.048 [2024-11-20 13:52:51.905904] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:32:45.048 [2024-11-20 13:52:51.906048] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:32:45.048 [2024-11-20 13:52:51.908046] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:45.048 "name": "Existed_Raid", 00:32:45.048 "uuid": "23094698-1de6-4290-a77a-215277b3c13f", 00:32:45.048 "strip_size_kb": 64, 00:32:45.048 "state": "configuring", 00:32:45.048 "raid_level": "raid5f", 00:32:45.048 "superblock": true, 00:32:45.048 "num_base_bdevs": 3, 00:32:45.048 "num_base_bdevs_discovered": 2, 00:32:45.048 "num_base_bdevs_operational": 3, 00:32:45.048 "base_bdevs_list": [ 00:32:45.048 { 00:32:45.048 "name": "BaseBdev1", 00:32:45.048 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:45.048 "is_configured": false, 00:32:45.048 "data_offset": 0, 00:32:45.048 "data_size": 0 00:32:45.048 }, 00:32:45.048 { 00:32:45.048 "name": "BaseBdev2", 00:32:45.048 "uuid": "1a411ecd-5151-427a-b2e7-ce9b4d0c40bb", 00:32:45.048 "is_configured": true, 00:32:45.048 "data_offset": 2048, 00:32:45.048 "data_size": 63488 00:32:45.048 }, 00:32:45.048 { 00:32:45.048 "name": "BaseBdev3", 00:32:45.048 "uuid": "cfc71633-cabf-47aa-92b3-71f72250cf01", 00:32:45.048 "is_configured": true, 00:32:45.048 "data_offset": 2048, 00:32:45.048 "data_size": 63488 00:32:45.048 } 00:32:45.048 ] 00:32:45.048 }' 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:45.048 13:52:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:45.306 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:32:45.306 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:45.306 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:45.306 [2024-11-20 13:52:52.221845] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:32:45.306 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:45.306 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:32:45.306 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:45.306 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:45.307 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:45.307 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:45.307 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:45.307 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:45.307 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:45.307 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:45.307 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:45.307 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:45.307 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:45.307 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:45.307 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:45.307 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:45.307 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:45.307 "name": "Existed_Raid", 00:32:45.307 "uuid": "23094698-1de6-4290-a77a-215277b3c13f", 00:32:45.307 "strip_size_kb": 64, 00:32:45.307 "state": "configuring", 00:32:45.307 "raid_level": "raid5f", 00:32:45.307 "superblock": true, 00:32:45.307 "num_base_bdevs": 3, 00:32:45.307 "num_base_bdevs_discovered": 1, 00:32:45.307 "num_base_bdevs_operational": 3, 00:32:45.307 "base_bdevs_list": [ 00:32:45.307 { 00:32:45.307 "name": "BaseBdev1", 00:32:45.307 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:45.307 "is_configured": false, 00:32:45.307 "data_offset": 0, 00:32:45.307 "data_size": 0 00:32:45.307 }, 00:32:45.307 { 00:32:45.307 "name": null, 00:32:45.307 "uuid": "1a411ecd-5151-427a-b2e7-ce9b4d0c40bb", 00:32:45.307 "is_configured": false, 00:32:45.307 "data_offset": 0, 00:32:45.307 "data_size": 63488 00:32:45.307 }, 00:32:45.307 { 00:32:45.307 "name": "BaseBdev3", 00:32:45.307 "uuid": "cfc71633-cabf-47aa-92b3-71f72250cf01", 00:32:45.307 "is_configured": true, 00:32:45.307 "data_offset": 2048, 00:32:45.307 "data_size": 63488 00:32:45.307 } 00:32:45.307 ] 00:32:45.307 }' 00:32:45.307 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:45.307 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:45.649 [2024-11-20 13:52:52.602143] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:32:45.649 BaseBdev1 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:45.649 [ 00:32:45.649 { 00:32:45.649 "name": "BaseBdev1", 00:32:45.649 "aliases": [ 00:32:45.649 "61f947a3-e37e-4456-834a-fd350a2c2f27" 00:32:45.649 ], 00:32:45.649 "product_name": "Malloc disk", 00:32:45.649 "block_size": 512, 00:32:45.649 "num_blocks": 65536, 00:32:45.649 "uuid": "61f947a3-e37e-4456-834a-fd350a2c2f27", 00:32:45.649 "assigned_rate_limits": { 00:32:45.649 "rw_ios_per_sec": 0, 00:32:45.649 "rw_mbytes_per_sec": 0, 00:32:45.649 "r_mbytes_per_sec": 0, 00:32:45.649 "w_mbytes_per_sec": 0 00:32:45.649 }, 00:32:45.649 "claimed": true, 00:32:45.649 "claim_type": "exclusive_write", 00:32:45.649 "zoned": false, 00:32:45.649 "supported_io_types": { 00:32:45.649 "read": true, 00:32:45.649 "write": true, 00:32:45.649 "unmap": true, 00:32:45.649 "flush": true, 00:32:45.649 "reset": true, 00:32:45.649 "nvme_admin": false, 00:32:45.649 "nvme_io": false, 00:32:45.649 "nvme_io_md": false, 00:32:45.649 "write_zeroes": true, 00:32:45.649 "zcopy": true, 00:32:45.649 "get_zone_info": false, 00:32:45.649 "zone_management": false, 00:32:45.649 "zone_append": false, 00:32:45.649 "compare": false, 00:32:45.649 "compare_and_write": false, 00:32:45.649 "abort": true, 00:32:45.649 "seek_hole": false, 00:32:45.649 "seek_data": false, 00:32:45.649 "copy": true, 00:32:45.649 "nvme_iov_md": false 00:32:45.649 }, 00:32:45.649 "memory_domains": [ 00:32:45.649 { 00:32:45.649 "dma_device_id": "system", 00:32:45.649 "dma_device_type": 1 00:32:45.649 }, 00:32:45.649 { 00:32:45.649 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:32:45.649 "dma_device_type": 2 00:32:45.649 } 00:32:45.649 ], 00:32:45.649 "driver_specific": {} 00:32:45.649 } 00:32:45.649 ] 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:45.649 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:45.650 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:45.650 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:45.650 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:45.650 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:45.650 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:45.650 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:45.650 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:45.650 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:45.650 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:45.650 "name": "Existed_Raid", 00:32:45.650 "uuid": "23094698-1de6-4290-a77a-215277b3c13f", 00:32:45.650 "strip_size_kb": 64, 00:32:45.650 "state": "configuring", 00:32:45.650 "raid_level": "raid5f", 00:32:45.650 "superblock": true, 00:32:45.650 "num_base_bdevs": 3, 00:32:45.650 "num_base_bdevs_discovered": 2, 00:32:45.650 "num_base_bdevs_operational": 3, 00:32:45.650 "base_bdevs_list": [ 00:32:45.650 { 00:32:45.650 "name": "BaseBdev1", 00:32:45.650 "uuid": "61f947a3-e37e-4456-834a-fd350a2c2f27", 00:32:45.650 "is_configured": true, 00:32:45.650 "data_offset": 2048, 00:32:45.650 "data_size": 63488 00:32:45.650 }, 00:32:45.650 { 00:32:45.650 "name": null, 00:32:45.650 "uuid": "1a411ecd-5151-427a-b2e7-ce9b4d0c40bb", 00:32:45.650 "is_configured": false, 00:32:45.650 "data_offset": 0, 00:32:45.650 "data_size": 63488 00:32:45.650 }, 00:32:45.650 { 00:32:45.650 "name": "BaseBdev3", 00:32:45.650 "uuid": "cfc71633-cabf-47aa-92b3-71f72250cf01", 00:32:45.650 "is_configured": true, 00:32:45.650 "data_offset": 2048, 00:32:45.650 "data_size": 63488 00:32:45.650 } 00:32:45.650 ] 00:32:45.650 }' 00:32:45.650 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:45.650 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:45.908 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:32:45.908 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:45.908 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:45.908 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:45.908 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:46.168 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:32:46.168 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:32:46.168 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:46.168 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:46.168 [2024-11-20 13:52:52.978268] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:32:46.168 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:46.168 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:32:46.168 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:46.168 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:46.168 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:46.168 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:46.168 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:46.168 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:46.168 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:46.168 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:46.168 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:46.168 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:46.168 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:46.168 13:52:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:46.168 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:46.168 13:52:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:46.168 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:46.168 "name": "Existed_Raid", 00:32:46.168 "uuid": "23094698-1de6-4290-a77a-215277b3c13f", 00:32:46.168 "strip_size_kb": 64, 00:32:46.168 "state": "configuring", 00:32:46.168 "raid_level": "raid5f", 00:32:46.168 "superblock": true, 00:32:46.168 "num_base_bdevs": 3, 00:32:46.168 "num_base_bdevs_discovered": 1, 00:32:46.168 "num_base_bdevs_operational": 3, 00:32:46.168 "base_bdevs_list": [ 00:32:46.168 { 00:32:46.168 "name": "BaseBdev1", 00:32:46.168 "uuid": "61f947a3-e37e-4456-834a-fd350a2c2f27", 00:32:46.168 "is_configured": true, 00:32:46.168 "data_offset": 2048, 00:32:46.168 "data_size": 63488 00:32:46.168 }, 00:32:46.168 { 00:32:46.168 "name": null, 00:32:46.168 "uuid": "1a411ecd-5151-427a-b2e7-ce9b4d0c40bb", 00:32:46.168 "is_configured": false, 00:32:46.168 "data_offset": 0, 00:32:46.168 "data_size": 63488 00:32:46.168 }, 00:32:46.168 { 00:32:46.168 "name": null, 00:32:46.168 "uuid": "cfc71633-cabf-47aa-92b3-71f72250cf01", 00:32:46.168 "is_configured": false, 00:32:46.168 "data_offset": 0, 00:32:46.168 "data_size": 63488 00:32:46.168 } 00:32:46.169 ] 00:32:46.169 }' 00:32:46.169 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:46.169 13:52:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:46.427 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:46.427 13:52:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:46.427 13:52:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:46.427 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:32:46.427 13:52:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:46.427 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:32:46.427 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:32:46.427 13:52:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:46.427 13:52:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:46.427 [2024-11-20 13:52:53.322384] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:32:46.427 13:52:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:46.427 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:32:46.427 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:46.427 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:46.427 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:46.427 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:46.427 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:46.427 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:46.427 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:46.427 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:46.427 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:46.428 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:46.428 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:46.428 13:52:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:46.428 13:52:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:46.428 13:52:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:46.428 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:46.428 "name": "Existed_Raid", 00:32:46.428 "uuid": "23094698-1de6-4290-a77a-215277b3c13f", 00:32:46.428 "strip_size_kb": 64, 00:32:46.428 "state": "configuring", 00:32:46.428 "raid_level": "raid5f", 00:32:46.428 "superblock": true, 00:32:46.428 "num_base_bdevs": 3, 00:32:46.428 "num_base_bdevs_discovered": 2, 00:32:46.428 "num_base_bdevs_operational": 3, 00:32:46.428 "base_bdevs_list": [ 00:32:46.428 { 00:32:46.428 "name": "BaseBdev1", 00:32:46.428 "uuid": "61f947a3-e37e-4456-834a-fd350a2c2f27", 00:32:46.428 "is_configured": true, 00:32:46.428 "data_offset": 2048, 00:32:46.428 "data_size": 63488 00:32:46.428 }, 00:32:46.428 { 00:32:46.428 "name": null, 00:32:46.428 "uuid": "1a411ecd-5151-427a-b2e7-ce9b4d0c40bb", 00:32:46.428 "is_configured": false, 00:32:46.428 "data_offset": 0, 00:32:46.428 "data_size": 63488 00:32:46.428 }, 00:32:46.428 { 00:32:46.428 "name": "BaseBdev3", 00:32:46.428 "uuid": "cfc71633-cabf-47aa-92b3-71f72250cf01", 00:32:46.428 "is_configured": true, 00:32:46.428 "data_offset": 2048, 00:32:46.428 "data_size": 63488 00:32:46.428 } 00:32:46.428 ] 00:32:46.428 }' 00:32:46.428 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:46.428 13:52:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:46.686 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:32:46.686 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:46.686 13:52:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:46.686 13:52:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:46.686 13:52:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:46.686 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:32:46.686 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:32:46.686 13:52:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:46.686 13:52:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:46.686 [2024-11-20 13:52:53.714484] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:32:46.945 13:52:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:46.945 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:32:46.945 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:46.945 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:46.945 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:46.945 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:46.945 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:46.945 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:46.945 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:46.945 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:46.945 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:46.945 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:46.945 13:52:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:46.945 13:52:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:46.945 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:46.945 13:52:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:46.945 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:46.945 "name": "Existed_Raid", 00:32:46.945 "uuid": "23094698-1de6-4290-a77a-215277b3c13f", 00:32:46.945 "strip_size_kb": 64, 00:32:46.945 "state": "configuring", 00:32:46.945 "raid_level": "raid5f", 00:32:46.945 "superblock": true, 00:32:46.945 "num_base_bdevs": 3, 00:32:46.945 "num_base_bdevs_discovered": 1, 00:32:46.945 "num_base_bdevs_operational": 3, 00:32:46.945 "base_bdevs_list": [ 00:32:46.945 { 00:32:46.945 "name": null, 00:32:46.945 "uuid": "61f947a3-e37e-4456-834a-fd350a2c2f27", 00:32:46.945 "is_configured": false, 00:32:46.945 "data_offset": 0, 00:32:46.945 "data_size": 63488 00:32:46.945 }, 00:32:46.945 { 00:32:46.945 "name": null, 00:32:46.945 "uuid": "1a411ecd-5151-427a-b2e7-ce9b4d0c40bb", 00:32:46.946 "is_configured": false, 00:32:46.946 "data_offset": 0, 00:32:46.946 "data_size": 63488 00:32:46.946 }, 00:32:46.946 { 00:32:46.946 "name": "BaseBdev3", 00:32:46.946 "uuid": "cfc71633-cabf-47aa-92b3-71f72250cf01", 00:32:46.946 "is_configured": true, 00:32:46.946 "data_offset": 2048, 00:32:46.946 "data_size": 63488 00:32:46.946 } 00:32:46.946 ] 00:32:46.946 }' 00:32:46.946 13:52:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:46.946 13:52:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:47.203 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:32:47.203 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:47.204 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:47.204 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:47.204 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:47.204 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:32:47.204 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:32:47.204 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:47.204 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:47.204 [2024-11-20 13:52:54.112040] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:32:47.204 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:47.204 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:32:47.204 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:47.204 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:47.204 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:47.204 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:47.204 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:47.204 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:47.204 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:47.204 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:47.204 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:47.204 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:47.204 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:47.204 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:47.204 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:47.204 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:47.204 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:47.204 "name": "Existed_Raid", 00:32:47.204 "uuid": "23094698-1de6-4290-a77a-215277b3c13f", 00:32:47.204 "strip_size_kb": 64, 00:32:47.204 "state": "configuring", 00:32:47.204 "raid_level": "raid5f", 00:32:47.204 "superblock": true, 00:32:47.204 "num_base_bdevs": 3, 00:32:47.204 "num_base_bdevs_discovered": 2, 00:32:47.204 "num_base_bdevs_operational": 3, 00:32:47.204 "base_bdevs_list": [ 00:32:47.204 { 00:32:47.204 "name": null, 00:32:47.204 "uuid": "61f947a3-e37e-4456-834a-fd350a2c2f27", 00:32:47.204 "is_configured": false, 00:32:47.204 "data_offset": 0, 00:32:47.204 "data_size": 63488 00:32:47.204 }, 00:32:47.204 { 00:32:47.204 "name": "BaseBdev2", 00:32:47.204 "uuid": "1a411ecd-5151-427a-b2e7-ce9b4d0c40bb", 00:32:47.204 "is_configured": true, 00:32:47.204 "data_offset": 2048, 00:32:47.204 "data_size": 63488 00:32:47.204 }, 00:32:47.204 { 00:32:47.204 "name": "BaseBdev3", 00:32:47.204 "uuid": "cfc71633-cabf-47aa-92b3-71f72250cf01", 00:32:47.204 "is_configured": true, 00:32:47.204 "data_offset": 2048, 00:32:47.204 "data_size": 63488 00:32:47.204 } 00:32:47.204 ] 00:32:47.204 }' 00:32:47.204 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:47.204 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:47.462 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:47.462 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:47.462 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:47.462 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:32:47.462 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:47.462 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:32:47.462 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:32:47.462 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:47.462 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:47.462 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:47.462 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:47.462 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 61f947a3-e37e-4456-834a-fd350a2c2f27 00:32:47.462 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:47.462 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:47.722 [2024-11-20 13:52:54.524477] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:32:47.722 NewBaseBdev 00:32:47.722 [2024-11-20 13:52:54.524847] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:32:47.722 [2024-11-20 13:52:54.524868] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:32:47.722 [2024-11-20 13:52:54.525129] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:47.722 [2024-11-20 13:52:54.528025] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:32:47.722 [2024-11-20 13:52:54.528041] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:32:47.722 [2024-11-20 13:52:54.528166] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:47.722 [ 00:32:47.722 { 00:32:47.722 "name": "NewBaseBdev", 00:32:47.722 "aliases": [ 00:32:47.722 "61f947a3-e37e-4456-834a-fd350a2c2f27" 00:32:47.722 ], 00:32:47.722 "product_name": "Malloc disk", 00:32:47.722 "block_size": 512, 00:32:47.722 "num_blocks": 65536, 00:32:47.722 "uuid": "61f947a3-e37e-4456-834a-fd350a2c2f27", 00:32:47.722 "assigned_rate_limits": { 00:32:47.722 "rw_ios_per_sec": 0, 00:32:47.722 "rw_mbytes_per_sec": 0, 00:32:47.722 "r_mbytes_per_sec": 0, 00:32:47.722 "w_mbytes_per_sec": 0 00:32:47.722 }, 00:32:47.722 "claimed": true, 00:32:47.722 "claim_type": "exclusive_write", 00:32:47.722 "zoned": false, 00:32:47.722 "supported_io_types": { 00:32:47.722 "read": true, 00:32:47.722 "write": true, 00:32:47.722 "unmap": true, 00:32:47.722 "flush": true, 00:32:47.722 "reset": true, 00:32:47.722 "nvme_admin": false, 00:32:47.722 "nvme_io": false, 00:32:47.722 "nvme_io_md": false, 00:32:47.722 "write_zeroes": true, 00:32:47.722 "zcopy": true, 00:32:47.722 "get_zone_info": false, 00:32:47.722 "zone_management": false, 00:32:47.722 "zone_append": false, 00:32:47.722 "compare": false, 00:32:47.722 "compare_and_write": false, 00:32:47.722 "abort": true, 00:32:47.722 "seek_hole": false, 00:32:47.722 "seek_data": false, 00:32:47.722 "copy": true, 00:32:47.722 "nvme_iov_md": false 00:32:47.722 }, 00:32:47.722 "memory_domains": [ 00:32:47.722 { 00:32:47.722 "dma_device_id": "system", 00:32:47.722 "dma_device_type": 1 00:32:47.722 }, 00:32:47.722 { 00:32:47.722 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:32:47.722 "dma_device_type": 2 00:32:47.722 } 00:32:47.722 ], 00:32:47.722 "driver_specific": {} 00:32:47.722 } 00:32:47.722 ] 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:47.722 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:47.722 "name": "Existed_Raid", 00:32:47.722 "uuid": "23094698-1de6-4290-a77a-215277b3c13f", 00:32:47.722 "strip_size_kb": 64, 00:32:47.722 "state": "online", 00:32:47.722 "raid_level": "raid5f", 00:32:47.722 "superblock": true, 00:32:47.722 "num_base_bdevs": 3, 00:32:47.722 "num_base_bdevs_discovered": 3, 00:32:47.723 "num_base_bdevs_operational": 3, 00:32:47.723 "base_bdevs_list": [ 00:32:47.723 { 00:32:47.723 "name": "NewBaseBdev", 00:32:47.723 "uuid": "61f947a3-e37e-4456-834a-fd350a2c2f27", 00:32:47.723 "is_configured": true, 00:32:47.723 "data_offset": 2048, 00:32:47.723 "data_size": 63488 00:32:47.723 }, 00:32:47.723 { 00:32:47.723 "name": "BaseBdev2", 00:32:47.723 "uuid": "1a411ecd-5151-427a-b2e7-ce9b4d0c40bb", 00:32:47.723 "is_configured": true, 00:32:47.723 "data_offset": 2048, 00:32:47.723 "data_size": 63488 00:32:47.723 }, 00:32:47.723 { 00:32:47.723 "name": "BaseBdev3", 00:32:47.723 "uuid": "cfc71633-cabf-47aa-92b3-71f72250cf01", 00:32:47.723 "is_configured": true, 00:32:47.723 "data_offset": 2048, 00:32:47.723 "data_size": 63488 00:32:47.723 } 00:32:47.723 ] 00:32:47.723 }' 00:32:47.723 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:47.723 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:47.988 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:32:47.988 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:32:47.988 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:32:47.988 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:32:47.988 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:32:47.988 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:32:47.988 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:32:47.988 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:32:47.988 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:47.988 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:47.988 [2024-11-20 13:52:54.876041] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:32:47.988 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:47.988 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:32:47.988 "name": "Existed_Raid", 00:32:47.988 "aliases": [ 00:32:47.988 "23094698-1de6-4290-a77a-215277b3c13f" 00:32:47.988 ], 00:32:47.988 "product_name": "Raid Volume", 00:32:47.988 "block_size": 512, 00:32:47.988 "num_blocks": 126976, 00:32:47.988 "uuid": "23094698-1de6-4290-a77a-215277b3c13f", 00:32:47.988 "assigned_rate_limits": { 00:32:47.988 "rw_ios_per_sec": 0, 00:32:47.988 "rw_mbytes_per_sec": 0, 00:32:47.988 "r_mbytes_per_sec": 0, 00:32:47.988 "w_mbytes_per_sec": 0 00:32:47.988 }, 00:32:47.988 "claimed": false, 00:32:47.988 "zoned": false, 00:32:47.988 "supported_io_types": { 00:32:47.988 "read": true, 00:32:47.988 "write": true, 00:32:47.988 "unmap": false, 00:32:47.988 "flush": false, 00:32:47.988 "reset": true, 00:32:47.988 "nvme_admin": false, 00:32:47.988 "nvme_io": false, 00:32:47.988 "nvme_io_md": false, 00:32:47.988 "write_zeroes": true, 00:32:47.988 "zcopy": false, 00:32:47.988 "get_zone_info": false, 00:32:47.988 "zone_management": false, 00:32:47.988 "zone_append": false, 00:32:47.988 "compare": false, 00:32:47.988 "compare_and_write": false, 00:32:47.988 "abort": false, 00:32:47.988 "seek_hole": false, 00:32:47.988 "seek_data": false, 00:32:47.988 "copy": false, 00:32:47.988 "nvme_iov_md": false 00:32:47.988 }, 00:32:47.988 "driver_specific": { 00:32:47.988 "raid": { 00:32:47.988 "uuid": "23094698-1de6-4290-a77a-215277b3c13f", 00:32:47.988 "strip_size_kb": 64, 00:32:47.988 "state": "online", 00:32:47.988 "raid_level": "raid5f", 00:32:47.988 "superblock": true, 00:32:47.988 "num_base_bdevs": 3, 00:32:47.988 "num_base_bdevs_discovered": 3, 00:32:47.988 "num_base_bdevs_operational": 3, 00:32:47.988 "base_bdevs_list": [ 00:32:47.988 { 00:32:47.988 "name": "NewBaseBdev", 00:32:47.988 "uuid": "61f947a3-e37e-4456-834a-fd350a2c2f27", 00:32:47.988 "is_configured": true, 00:32:47.988 "data_offset": 2048, 00:32:47.988 "data_size": 63488 00:32:47.988 }, 00:32:47.988 { 00:32:47.988 "name": "BaseBdev2", 00:32:47.988 "uuid": "1a411ecd-5151-427a-b2e7-ce9b4d0c40bb", 00:32:47.988 "is_configured": true, 00:32:47.988 "data_offset": 2048, 00:32:47.988 "data_size": 63488 00:32:47.988 }, 00:32:47.988 { 00:32:47.988 "name": "BaseBdev3", 00:32:47.988 "uuid": "cfc71633-cabf-47aa-92b3-71f72250cf01", 00:32:47.988 "is_configured": true, 00:32:47.988 "data_offset": 2048, 00:32:47.988 "data_size": 63488 00:32:47.988 } 00:32:47.988 ] 00:32:47.988 } 00:32:47.989 } 00:32:47.989 }' 00:32:47.989 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:32:47.989 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:32:47.989 BaseBdev2 00:32:47.989 BaseBdev3' 00:32:47.989 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:47.989 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:32:47.989 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:32:47.989 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:32:47.989 13:52:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:47.989 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:47.989 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:47.989 13:52:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:47.989 13:52:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:32:47.989 13:52:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:32:47.989 13:52:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:32:47.989 13:52:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:32:47.989 13:52:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:47.989 13:52:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:47.989 13:52:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:47.989 13:52:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:47.989 13:52:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:32:47.989 13:52:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:32:47.989 13:52:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:32:47.989 13:52:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:32:47.989 13:52:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:47.989 13:52:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:47.989 13:52:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:48.247 13:52:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:48.247 13:52:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:32:48.247 13:52:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:32:48.247 13:52:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:32:48.247 13:52:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:48.247 13:52:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:48.247 [2024-11-20 13:52:55.075864] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:32:48.247 [2024-11-20 13:52:55.075963] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:32:48.247 [2024-11-20 13:52:55.076054] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:32:48.247 [2024-11-20 13:52:55.076301] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:32:48.247 [2024-11-20 13:52:55.076312] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:32:48.247 13:52:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:48.247 13:52:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 78232 00:32:48.247 13:52:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 78232 ']' 00:32:48.247 13:52:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 78232 00:32:48.247 13:52:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:32:48.247 13:52:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:32:48.247 13:52:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 78232 00:32:48.247 killing process with pid 78232 00:32:48.247 13:52:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:32:48.247 13:52:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:32:48.247 13:52:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 78232' 00:32:48.247 13:52:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 78232 00:32:48.247 [2024-11-20 13:52:55.106951] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:32:48.247 13:52:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 78232 00:32:48.247 [2024-11-20 13:52:55.264371] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:32:49.182 13:52:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:32:49.182 00:32:49.182 real 0m7.551s 00:32:49.182 user 0m11.970s 00:32:49.182 sys 0m1.351s 00:32:49.182 13:52:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:32:49.182 13:52:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:32:49.182 ************************************ 00:32:49.182 END TEST raid5f_state_function_test_sb 00:32:49.182 ************************************ 00:32:49.182 13:52:55 bdev_raid -- bdev/bdev_raid.sh@988 -- # run_test raid5f_superblock_test raid_superblock_test raid5f 3 00:32:49.182 13:52:55 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:32:49.182 13:52:55 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:32:49.182 13:52:55 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:32:49.182 ************************************ 00:32:49.182 START TEST raid5f_superblock_test 00:32:49.182 ************************************ 00:32:49.182 13:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid5f 3 00:32:49.182 13:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid5f 00:32:49.182 13:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:32:49.182 13:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:32:49.182 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:49.182 13:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:32:49.182 13:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:32:49.182 13:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:32:49.182 13:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:32:49.182 13:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:32:49.182 13:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:32:49.182 13:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:32:49.182 13:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:32:49.182 13:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:32:49.182 13:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:32:49.182 13:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid5f '!=' raid1 ']' 00:32:49.182 13:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:32:49.182 13:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:32:49.182 13:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=78819 00:32:49.182 13:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 78819 00:32:49.182 13:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 78819 ']' 00:32:49.182 13:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:49.182 13:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:32:49.182 13:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:49.182 13:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:32:49.182 13:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:49.182 13:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:32:49.182 [2024-11-20 13:52:56.001291] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:32:49.183 [2024-11-20 13:52:56.001436] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid78819 ] 00:32:49.183 [2024-11-20 13:52:56.155497] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:49.441 [2024-11-20 13:52:56.275017] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:32:49.441 [2024-11-20 13:52:56.423345] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:32:49.441 [2024-11-20 13:52:56.423404] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:32:50.007 13:52:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:50.008 malloc1 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:50.008 [2024-11-20 13:52:56.891531] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:32:50.008 [2024-11-20 13:52:56.891599] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:50.008 [2024-11-20 13:52:56.891623] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:32:50.008 [2024-11-20 13:52:56.891633] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:50.008 [2024-11-20 13:52:56.893935] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:50.008 [2024-11-20 13:52:56.893974] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:32:50.008 pt1 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:50.008 malloc2 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:50.008 [2024-11-20 13:52:56.929745] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:32:50.008 [2024-11-20 13:52:56.929793] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:50.008 [2024-11-20 13:52:56.929818] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:32:50.008 [2024-11-20 13:52:56.929827] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:50.008 [2024-11-20 13:52:56.932053] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:50.008 [2024-11-20 13:52:56.932086] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:32:50.008 pt2 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:50.008 malloc3 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:50.008 [2024-11-20 13:52:56.991053] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:32:50.008 [2024-11-20 13:52:56.991235] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:50.008 [2024-11-20 13:52:56.991282] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:32:50.008 [2024-11-20 13:52:56.991636] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:50.008 [2024-11-20 13:52:56.994020] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:50.008 [2024-11-20 13:52:56.994138] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:32:50.008 pt3 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:50.008 13:52:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:50.008 [2024-11-20 13:52:56.999081] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:32:50.008 [2024-11-20 13:52:57.001124] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:32:50.008 [2024-11-20 13:52:57.001269] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:32:50.008 [2024-11-20 13:52:57.001442] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:32:50.008 [2024-11-20 13:52:57.001461] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:32:50.008 [2024-11-20 13:52:57.001719] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:32:50.008 [2024-11-20 13:52:57.005556] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:32:50.008 [2024-11-20 13:52:57.005575] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:32:50.008 [2024-11-20 13:52:57.005758] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:50.008 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:50.008 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:32:50.008 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:50.008 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:50.008 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:50.008 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:50.008 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:50.008 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:50.008 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:50.008 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:50.008 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:50.008 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:50.008 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:50.008 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:50.008 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:50.008 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:50.008 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:50.008 "name": "raid_bdev1", 00:32:50.008 "uuid": "9628b010-45c1-49e2-b5cd-96833ff4c37d", 00:32:50.008 "strip_size_kb": 64, 00:32:50.008 "state": "online", 00:32:50.008 "raid_level": "raid5f", 00:32:50.008 "superblock": true, 00:32:50.008 "num_base_bdevs": 3, 00:32:50.008 "num_base_bdevs_discovered": 3, 00:32:50.008 "num_base_bdevs_operational": 3, 00:32:50.008 "base_bdevs_list": [ 00:32:50.008 { 00:32:50.008 "name": "pt1", 00:32:50.008 "uuid": "00000000-0000-0000-0000-000000000001", 00:32:50.008 "is_configured": true, 00:32:50.008 "data_offset": 2048, 00:32:50.008 "data_size": 63488 00:32:50.008 }, 00:32:50.008 { 00:32:50.008 "name": "pt2", 00:32:50.009 "uuid": "00000000-0000-0000-0000-000000000002", 00:32:50.009 "is_configured": true, 00:32:50.009 "data_offset": 2048, 00:32:50.009 "data_size": 63488 00:32:50.009 }, 00:32:50.009 { 00:32:50.009 "name": "pt3", 00:32:50.009 "uuid": "00000000-0000-0000-0000-000000000003", 00:32:50.009 "is_configured": true, 00:32:50.009 "data_offset": 2048, 00:32:50.009 "data_size": 63488 00:32:50.009 } 00:32:50.009 ] 00:32:50.009 }' 00:32:50.009 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:50.009 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:50.266 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:32:50.266 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:32:50.266 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:32:50.266 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:32:50.266 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:32:50.266 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:32:50.266 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:32:50.266 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:32:50.266 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:50.266 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:50.266 [2024-11-20 13:52:57.314331] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:32:50.266 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:50.525 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:32:50.525 "name": "raid_bdev1", 00:32:50.525 "aliases": [ 00:32:50.525 "9628b010-45c1-49e2-b5cd-96833ff4c37d" 00:32:50.525 ], 00:32:50.525 "product_name": "Raid Volume", 00:32:50.525 "block_size": 512, 00:32:50.525 "num_blocks": 126976, 00:32:50.525 "uuid": "9628b010-45c1-49e2-b5cd-96833ff4c37d", 00:32:50.525 "assigned_rate_limits": { 00:32:50.525 "rw_ios_per_sec": 0, 00:32:50.525 "rw_mbytes_per_sec": 0, 00:32:50.525 "r_mbytes_per_sec": 0, 00:32:50.525 "w_mbytes_per_sec": 0 00:32:50.525 }, 00:32:50.525 "claimed": false, 00:32:50.525 "zoned": false, 00:32:50.525 "supported_io_types": { 00:32:50.525 "read": true, 00:32:50.525 "write": true, 00:32:50.525 "unmap": false, 00:32:50.525 "flush": false, 00:32:50.525 "reset": true, 00:32:50.525 "nvme_admin": false, 00:32:50.525 "nvme_io": false, 00:32:50.525 "nvme_io_md": false, 00:32:50.525 "write_zeroes": true, 00:32:50.525 "zcopy": false, 00:32:50.526 "get_zone_info": false, 00:32:50.526 "zone_management": false, 00:32:50.526 "zone_append": false, 00:32:50.526 "compare": false, 00:32:50.526 "compare_and_write": false, 00:32:50.526 "abort": false, 00:32:50.526 "seek_hole": false, 00:32:50.526 "seek_data": false, 00:32:50.526 "copy": false, 00:32:50.526 "nvme_iov_md": false 00:32:50.526 }, 00:32:50.526 "driver_specific": { 00:32:50.526 "raid": { 00:32:50.526 "uuid": "9628b010-45c1-49e2-b5cd-96833ff4c37d", 00:32:50.526 "strip_size_kb": 64, 00:32:50.526 "state": "online", 00:32:50.526 "raid_level": "raid5f", 00:32:50.526 "superblock": true, 00:32:50.526 "num_base_bdevs": 3, 00:32:50.526 "num_base_bdevs_discovered": 3, 00:32:50.526 "num_base_bdevs_operational": 3, 00:32:50.526 "base_bdevs_list": [ 00:32:50.526 { 00:32:50.526 "name": "pt1", 00:32:50.526 "uuid": "00000000-0000-0000-0000-000000000001", 00:32:50.526 "is_configured": true, 00:32:50.526 "data_offset": 2048, 00:32:50.526 "data_size": 63488 00:32:50.526 }, 00:32:50.526 { 00:32:50.526 "name": "pt2", 00:32:50.526 "uuid": "00000000-0000-0000-0000-000000000002", 00:32:50.526 "is_configured": true, 00:32:50.526 "data_offset": 2048, 00:32:50.526 "data_size": 63488 00:32:50.526 }, 00:32:50.526 { 00:32:50.526 "name": "pt3", 00:32:50.526 "uuid": "00000000-0000-0000-0000-000000000003", 00:32:50.526 "is_configured": true, 00:32:50.526 "data_offset": 2048, 00:32:50.526 "data_size": 63488 00:32:50.526 } 00:32:50.526 ] 00:32:50.526 } 00:32:50.526 } 00:32:50.526 }' 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:32:50.526 pt2 00:32:50.526 pt3' 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:50.526 [2024-11-20 13:52:57.522328] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=9628b010-45c1-49e2-b5cd-96833ff4c37d 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 9628b010-45c1-49e2-b5cd-96833ff4c37d ']' 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:50.526 [2024-11-20 13:52:57.550138] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:32:50.526 [2024-11-20 13:52:57.550168] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:32:50.526 [2024-11-20 13:52:57.550256] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:32:50.526 [2024-11-20 13:52:57.550343] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:32:50.526 [2024-11-20 13:52:57.550353] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:50.526 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:32:50.785 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:50.786 [2024-11-20 13:52:57.654220] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:32:50.786 [2024-11-20 13:52:57.656275] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:32:50.786 [2024-11-20 13:52:57.656438] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:32:50.786 [2024-11-20 13:52:57.656504] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:32:50.786 [2024-11-20 13:52:57.656562] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:32:50.786 [2024-11-20 13:52:57.656582] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:32:50.786 [2024-11-20 13:52:57.656599] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:32:50.786 [2024-11-20 13:52:57.656609] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:32:50.786 request: 00:32:50.786 { 00:32:50.786 "name": "raid_bdev1", 00:32:50.786 "raid_level": "raid5f", 00:32:50.786 "base_bdevs": [ 00:32:50.786 "malloc1", 00:32:50.786 "malloc2", 00:32:50.786 "malloc3" 00:32:50.786 ], 00:32:50.786 "strip_size_kb": 64, 00:32:50.786 "superblock": false, 00:32:50.786 "method": "bdev_raid_create", 00:32:50.786 "req_id": 1 00:32:50.786 } 00:32:50.786 Got JSON-RPC error response 00:32:50.786 response: 00:32:50.786 { 00:32:50.786 "code": -17, 00:32:50.786 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:32:50.786 } 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:50.786 [2024-11-20 13:52:57.698185] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:32:50.786 [2024-11-20 13:52:57.698248] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:50.786 [2024-11-20 13:52:57.698269] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:32:50.786 [2024-11-20 13:52:57.698278] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:50.786 [2024-11-20 13:52:57.700659] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:50.786 [2024-11-20 13:52:57.700692] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:32:50.786 [2024-11-20 13:52:57.700778] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:32:50.786 [2024-11-20 13:52:57.700830] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:32:50.786 pt1 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:50.786 "name": "raid_bdev1", 00:32:50.786 "uuid": "9628b010-45c1-49e2-b5cd-96833ff4c37d", 00:32:50.786 "strip_size_kb": 64, 00:32:50.786 "state": "configuring", 00:32:50.786 "raid_level": "raid5f", 00:32:50.786 "superblock": true, 00:32:50.786 "num_base_bdevs": 3, 00:32:50.786 "num_base_bdevs_discovered": 1, 00:32:50.786 "num_base_bdevs_operational": 3, 00:32:50.786 "base_bdevs_list": [ 00:32:50.786 { 00:32:50.786 "name": "pt1", 00:32:50.786 "uuid": "00000000-0000-0000-0000-000000000001", 00:32:50.786 "is_configured": true, 00:32:50.786 "data_offset": 2048, 00:32:50.786 "data_size": 63488 00:32:50.786 }, 00:32:50.786 { 00:32:50.786 "name": null, 00:32:50.786 "uuid": "00000000-0000-0000-0000-000000000002", 00:32:50.786 "is_configured": false, 00:32:50.786 "data_offset": 2048, 00:32:50.786 "data_size": 63488 00:32:50.786 }, 00:32:50.786 { 00:32:50.786 "name": null, 00:32:50.786 "uuid": "00000000-0000-0000-0000-000000000003", 00:32:50.786 "is_configured": false, 00:32:50.786 "data_offset": 2048, 00:32:50.786 "data_size": 63488 00:32:50.786 } 00:32:50.786 ] 00:32:50.786 }' 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:50.786 13:52:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:51.044 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:32:51.044 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:32:51.044 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:51.044 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:51.044 [2024-11-20 13:52:58.010282] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:32:51.044 [2024-11-20 13:52:58.010355] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:51.044 [2024-11-20 13:52:58.010379] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:32:51.044 [2024-11-20 13:52:58.010389] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:51.044 [2024-11-20 13:52:58.010857] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:51.044 [2024-11-20 13:52:58.010877] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:32:51.044 [2024-11-20 13:52:58.010967] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:32:51.044 [2024-11-20 13:52:58.011008] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:32:51.044 pt2 00:32:51.044 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:51.044 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:32:51.044 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:51.044 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:51.044 [2024-11-20 13:52:58.018333] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:32:51.044 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:51.044 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:32:51.044 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:51.044 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:51.044 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:51.044 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:51.044 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:51.044 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:51.045 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:51.045 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:51.045 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:51.045 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:51.045 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:51.045 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:51.045 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:51.045 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:51.045 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:51.045 "name": "raid_bdev1", 00:32:51.045 "uuid": "9628b010-45c1-49e2-b5cd-96833ff4c37d", 00:32:51.045 "strip_size_kb": 64, 00:32:51.045 "state": "configuring", 00:32:51.045 "raid_level": "raid5f", 00:32:51.045 "superblock": true, 00:32:51.045 "num_base_bdevs": 3, 00:32:51.045 "num_base_bdevs_discovered": 1, 00:32:51.045 "num_base_bdevs_operational": 3, 00:32:51.045 "base_bdevs_list": [ 00:32:51.045 { 00:32:51.045 "name": "pt1", 00:32:51.045 "uuid": "00000000-0000-0000-0000-000000000001", 00:32:51.045 "is_configured": true, 00:32:51.045 "data_offset": 2048, 00:32:51.045 "data_size": 63488 00:32:51.045 }, 00:32:51.045 { 00:32:51.045 "name": null, 00:32:51.045 "uuid": "00000000-0000-0000-0000-000000000002", 00:32:51.045 "is_configured": false, 00:32:51.045 "data_offset": 0, 00:32:51.045 "data_size": 63488 00:32:51.045 }, 00:32:51.045 { 00:32:51.045 "name": null, 00:32:51.045 "uuid": "00000000-0000-0000-0000-000000000003", 00:32:51.045 "is_configured": false, 00:32:51.045 "data_offset": 2048, 00:32:51.045 "data_size": 63488 00:32:51.045 } 00:32:51.045 ] 00:32:51.045 }' 00:32:51.045 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:51.045 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:51.303 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:32:51.303 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:32:51.303 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:32:51.303 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:51.303 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:51.303 [2024-11-20 13:52:58.354358] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:32:51.303 [2024-11-20 13:52:58.354445] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:51.303 [2024-11-20 13:52:58.354467] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:32:51.303 [2024-11-20 13:52:58.354480] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:51.303 [2024-11-20 13:52:58.355011] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:51.303 [2024-11-20 13:52:58.355032] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:32:51.303 [2024-11-20 13:52:58.355125] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:32:51.303 [2024-11-20 13:52:58.355150] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:32:51.303 pt2 00:32:51.303 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:51.303 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:32:51.303 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:32:51.303 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:32:51.303 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:51.303 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:51.562 [2024-11-20 13:52:58.362339] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:32:51.562 [2024-11-20 13:52:58.362391] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:51.562 [2024-11-20 13:52:58.362408] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:32:51.562 [2024-11-20 13:52:58.362419] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:51.562 [2024-11-20 13:52:58.362855] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:51.562 [2024-11-20 13:52:58.362878] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:32:51.562 [2024-11-20 13:52:58.362950] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:32:51.562 [2024-11-20 13:52:58.362973] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:32:51.562 [2024-11-20 13:52:58.363119] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:32:51.562 [2024-11-20 13:52:58.363130] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:32:51.562 [2024-11-20 13:52:58.363373] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:32:51.562 [2024-11-20 13:52:58.367065] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:32:51.562 [2024-11-20 13:52:58.367084] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:32:51.562 [2024-11-20 13:52:58.367275] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:51.562 pt3 00:32:51.562 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:51.562 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:32:51.562 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:32:51.563 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:32:51.563 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:51.563 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:51.563 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:51.563 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:51.563 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:51.563 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:51.563 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:51.563 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:51.563 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:51.563 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:51.563 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:51.563 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:51.563 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:51.563 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:51.563 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:51.563 "name": "raid_bdev1", 00:32:51.563 "uuid": "9628b010-45c1-49e2-b5cd-96833ff4c37d", 00:32:51.563 "strip_size_kb": 64, 00:32:51.563 "state": "online", 00:32:51.563 "raid_level": "raid5f", 00:32:51.563 "superblock": true, 00:32:51.563 "num_base_bdevs": 3, 00:32:51.563 "num_base_bdevs_discovered": 3, 00:32:51.563 "num_base_bdevs_operational": 3, 00:32:51.563 "base_bdevs_list": [ 00:32:51.563 { 00:32:51.563 "name": "pt1", 00:32:51.563 "uuid": "00000000-0000-0000-0000-000000000001", 00:32:51.563 "is_configured": true, 00:32:51.563 "data_offset": 2048, 00:32:51.563 "data_size": 63488 00:32:51.563 }, 00:32:51.563 { 00:32:51.563 "name": "pt2", 00:32:51.563 "uuid": "00000000-0000-0000-0000-000000000002", 00:32:51.563 "is_configured": true, 00:32:51.563 "data_offset": 2048, 00:32:51.563 "data_size": 63488 00:32:51.563 }, 00:32:51.563 { 00:32:51.563 "name": "pt3", 00:32:51.563 "uuid": "00000000-0000-0000-0000-000000000003", 00:32:51.563 "is_configured": true, 00:32:51.563 "data_offset": 2048, 00:32:51.563 "data_size": 63488 00:32:51.563 } 00:32:51.563 ] 00:32:51.563 }' 00:32:51.563 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:51.563 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:51.821 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:32:51.821 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:32:51.821 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:32:51.821 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:32:51.821 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:32:51.821 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:32:51.821 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:32:51.821 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:32:51.821 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:51.821 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:51.821 [2024-11-20 13:52:58.747856] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:32:51.821 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:51.821 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:32:51.821 "name": "raid_bdev1", 00:32:51.821 "aliases": [ 00:32:51.821 "9628b010-45c1-49e2-b5cd-96833ff4c37d" 00:32:51.821 ], 00:32:51.821 "product_name": "Raid Volume", 00:32:51.821 "block_size": 512, 00:32:51.821 "num_blocks": 126976, 00:32:51.821 "uuid": "9628b010-45c1-49e2-b5cd-96833ff4c37d", 00:32:51.821 "assigned_rate_limits": { 00:32:51.821 "rw_ios_per_sec": 0, 00:32:51.821 "rw_mbytes_per_sec": 0, 00:32:51.821 "r_mbytes_per_sec": 0, 00:32:51.821 "w_mbytes_per_sec": 0 00:32:51.821 }, 00:32:51.821 "claimed": false, 00:32:51.821 "zoned": false, 00:32:51.821 "supported_io_types": { 00:32:51.821 "read": true, 00:32:51.821 "write": true, 00:32:51.821 "unmap": false, 00:32:51.821 "flush": false, 00:32:51.821 "reset": true, 00:32:51.821 "nvme_admin": false, 00:32:51.821 "nvme_io": false, 00:32:51.821 "nvme_io_md": false, 00:32:51.821 "write_zeroes": true, 00:32:51.821 "zcopy": false, 00:32:51.821 "get_zone_info": false, 00:32:51.821 "zone_management": false, 00:32:51.821 "zone_append": false, 00:32:51.821 "compare": false, 00:32:51.821 "compare_and_write": false, 00:32:51.821 "abort": false, 00:32:51.821 "seek_hole": false, 00:32:51.821 "seek_data": false, 00:32:51.821 "copy": false, 00:32:51.821 "nvme_iov_md": false 00:32:51.821 }, 00:32:51.821 "driver_specific": { 00:32:51.821 "raid": { 00:32:51.821 "uuid": "9628b010-45c1-49e2-b5cd-96833ff4c37d", 00:32:51.821 "strip_size_kb": 64, 00:32:51.821 "state": "online", 00:32:51.821 "raid_level": "raid5f", 00:32:51.821 "superblock": true, 00:32:51.821 "num_base_bdevs": 3, 00:32:51.821 "num_base_bdevs_discovered": 3, 00:32:51.821 "num_base_bdevs_operational": 3, 00:32:51.821 "base_bdevs_list": [ 00:32:51.821 { 00:32:51.821 "name": "pt1", 00:32:51.821 "uuid": "00000000-0000-0000-0000-000000000001", 00:32:51.821 "is_configured": true, 00:32:51.821 "data_offset": 2048, 00:32:51.821 "data_size": 63488 00:32:51.821 }, 00:32:51.821 { 00:32:51.822 "name": "pt2", 00:32:51.822 "uuid": "00000000-0000-0000-0000-000000000002", 00:32:51.822 "is_configured": true, 00:32:51.822 "data_offset": 2048, 00:32:51.822 "data_size": 63488 00:32:51.822 }, 00:32:51.822 { 00:32:51.822 "name": "pt3", 00:32:51.822 "uuid": "00000000-0000-0000-0000-000000000003", 00:32:51.822 "is_configured": true, 00:32:51.822 "data_offset": 2048, 00:32:51.822 "data_size": 63488 00:32:51.822 } 00:32:51.822 ] 00:32:51.822 } 00:32:51.822 } 00:32:51.822 }' 00:32:51.822 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:32:51.822 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:32:51.822 pt2 00:32:51.822 pt3' 00:32:51.822 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:51.822 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:32:51.822 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:32:51.822 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:32:51.822 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:51.822 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:51.822 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:51.822 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:51.822 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:32:51.822 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:32:51.822 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:52.080 [2024-11-20 13:52:58.955900] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 9628b010-45c1-49e2-b5cd-96833ff4c37d '!=' 9628b010-45c1-49e2-b5cd-96833ff4c37d ']' 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid5f 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:52.080 [2024-11-20 13:52:58.987731] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:52.080 13:52:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:52.080 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:52.080 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:52.080 "name": "raid_bdev1", 00:32:52.080 "uuid": "9628b010-45c1-49e2-b5cd-96833ff4c37d", 00:32:52.080 "strip_size_kb": 64, 00:32:52.080 "state": "online", 00:32:52.080 "raid_level": "raid5f", 00:32:52.080 "superblock": true, 00:32:52.080 "num_base_bdevs": 3, 00:32:52.080 "num_base_bdevs_discovered": 2, 00:32:52.080 "num_base_bdevs_operational": 2, 00:32:52.080 "base_bdevs_list": [ 00:32:52.080 { 00:32:52.080 "name": null, 00:32:52.080 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:52.081 "is_configured": false, 00:32:52.081 "data_offset": 0, 00:32:52.081 "data_size": 63488 00:32:52.081 }, 00:32:52.081 { 00:32:52.081 "name": "pt2", 00:32:52.081 "uuid": "00000000-0000-0000-0000-000000000002", 00:32:52.081 "is_configured": true, 00:32:52.081 "data_offset": 2048, 00:32:52.081 "data_size": 63488 00:32:52.081 }, 00:32:52.081 { 00:32:52.081 "name": "pt3", 00:32:52.081 "uuid": "00000000-0000-0000-0000-000000000003", 00:32:52.081 "is_configured": true, 00:32:52.081 "data_offset": 2048, 00:32:52.081 "data_size": 63488 00:32:52.081 } 00:32:52.081 ] 00:32:52.081 }' 00:32:52.081 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:52.081 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:52.344 [2024-11-20 13:52:59.299764] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:32:52.344 [2024-11-20 13:52:59.299797] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:32:52.344 [2024-11-20 13:52:59.299881] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:32:52.344 [2024-11-20 13:52:59.299950] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:32:52.344 [2024-11-20 13:52:59.299964] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:52.344 [2024-11-20 13:52:59.355729] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:32:52.344 [2024-11-20 13:52:59.355791] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:52.344 [2024-11-20 13:52:59.355809] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:32:52.344 [2024-11-20 13:52:59.355820] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:52.344 [2024-11-20 13:52:59.358488] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:52.344 [2024-11-20 13:52:59.358531] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:32:52.344 [2024-11-20 13:52:59.358621] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:32:52.344 [2024-11-20 13:52:59.358672] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:32:52.344 pt2 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 2 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:52.344 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:52.612 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:52.612 "name": "raid_bdev1", 00:32:52.612 "uuid": "9628b010-45c1-49e2-b5cd-96833ff4c37d", 00:32:52.612 "strip_size_kb": 64, 00:32:52.612 "state": "configuring", 00:32:52.612 "raid_level": "raid5f", 00:32:52.612 "superblock": true, 00:32:52.612 "num_base_bdevs": 3, 00:32:52.612 "num_base_bdevs_discovered": 1, 00:32:52.612 "num_base_bdevs_operational": 2, 00:32:52.612 "base_bdevs_list": [ 00:32:52.612 { 00:32:52.612 "name": null, 00:32:52.612 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:52.612 "is_configured": false, 00:32:52.612 "data_offset": 2048, 00:32:52.612 "data_size": 63488 00:32:52.612 }, 00:32:52.612 { 00:32:52.612 "name": "pt2", 00:32:52.612 "uuid": "00000000-0000-0000-0000-000000000002", 00:32:52.612 "is_configured": true, 00:32:52.612 "data_offset": 2048, 00:32:52.612 "data_size": 63488 00:32:52.612 }, 00:32:52.612 { 00:32:52.612 "name": null, 00:32:52.612 "uuid": "00000000-0000-0000-0000-000000000003", 00:32:52.612 "is_configured": false, 00:32:52.612 "data_offset": 2048, 00:32:52.612 "data_size": 63488 00:32:52.612 } 00:32:52.612 ] 00:32:52.612 }' 00:32:52.612 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:52.612 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:52.870 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:32:52.870 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:32:52.870 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@519 -- # i=2 00:32:52.870 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:32:52.870 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:52.870 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:52.870 [2024-11-20 13:52:59.691822] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:32:52.870 [2024-11-20 13:52:59.691900] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:52.870 [2024-11-20 13:52:59.691920] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:32:52.870 [2024-11-20 13:52:59.691931] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:52.870 [2024-11-20 13:52:59.692396] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:52.870 [2024-11-20 13:52:59.692418] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:32:52.870 [2024-11-20 13:52:59.692496] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:32:52.870 [2024-11-20 13:52:59.692522] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:32:52.871 [2024-11-20 13:52:59.692645] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:32:52.871 [2024-11-20 13:52:59.692656] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:32:52.871 [2024-11-20 13:52:59.692874] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:32:52.871 [2024-11-20 13:52:59.695749] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:32:52.871 [2024-11-20 13:52:59.695767] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:32:52.871 [2024-11-20 13:52:59.696019] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:52.871 pt3 00:32:52.871 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:52.871 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:32:52.871 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:52.871 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:52.871 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:52.871 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:52.871 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:52.871 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:52.871 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:52.871 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:52.871 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:52.871 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:52.871 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:52.871 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:52.871 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:52.871 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:52.871 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:52.871 "name": "raid_bdev1", 00:32:52.871 "uuid": "9628b010-45c1-49e2-b5cd-96833ff4c37d", 00:32:52.871 "strip_size_kb": 64, 00:32:52.871 "state": "online", 00:32:52.871 "raid_level": "raid5f", 00:32:52.871 "superblock": true, 00:32:52.871 "num_base_bdevs": 3, 00:32:52.871 "num_base_bdevs_discovered": 2, 00:32:52.871 "num_base_bdevs_operational": 2, 00:32:52.871 "base_bdevs_list": [ 00:32:52.871 { 00:32:52.871 "name": null, 00:32:52.871 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:52.871 "is_configured": false, 00:32:52.871 "data_offset": 2048, 00:32:52.871 "data_size": 63488 00:32:52.871 }, 00:32:52.871 { 00:32:52.871 "name": "pt2", 00:32:52.871 "uuid": "00000000-0000-0000-0000-000000000002", 00:32:52.871 "is_configured": true, 00:32:52.871 "data_offset": 2048, 00:32:52.871 "data_size": 63488 00:32:52.871 }, 00:32:52.871 { 00:32:52.871 "name": "pt3", 00:32:52.871 "uuid": "00000000-0000-0000-0000-000000000003", 00:32:52.871 "is_configured": true, 00:32:52.871 "data_offset": 2048, 00:32:52.871 "data_size": 63488 00:32:52.871 } 00:32:52.871 ] 00:32:52.871 }' 00:32:52.871 13:52:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:52.871 13:52:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:53.129 [2024-11-20 13:53:00.011940] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:32:53.129 [2024-11-20 13:53:00.011971] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:32:53.129 [2024-11-20 13:53:00.012059] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:32:53.129 [2024-11-20 13:53:00.012124] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:32:53.129 [2024-11-20 13:53:00.012133] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 3 -gt 2 ']' 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@534 -- # i=2 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt3 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:53.129 [2024-11-20 13:53:00.059974] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:32:53.129 [2024-11-20 13:53:00.060044] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:53.129 [2024-11-20 13:53:00.060063] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:32:53.129 [2024-11-20 13:53:00.060071] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:53.129 [2024-11-20 13:53:00.062174] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:53.129 [2024-11-20 13:53:00.062206] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:32:53.129 [2024-11-20 13:53:00.062286] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:32:53.129 [2024-11-20 13:53:00.062329] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:32:53.129 [2024-11-20 13:53:00.062447] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:32:53.129 [2024-11-20 13:53:00.062457] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:32:53.129 [2024-11-20 13:53:00.062473] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:32:53.129 [2024-11-20 13:53:00.062513] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:32:53.129 pt1 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 3 -gt 2 ']' 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 2 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:53.129 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:53.130 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:53.130 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:53.130 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:53.130 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:53.130 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:53.130 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:53.130 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:53.130 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:53.130 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:53.130 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:53.130 "name": "raid_bdev1", 00:32:53.130 "uuid": "9628b010-45c1-49e2-b5cd-96833ff4c37d", 00:32:53.130 "strip_size_kb": 64, 00:32:53.130 "state": "configuring", 00:32:53.130 "raid_level": "raid5f", 00:32:53.130 "superblock": true, 00:32:53.130 "num_base_bdevs": 3, 00:32:53.130 "num_base_bdevs_discovered": 1, 00:32:53.130 "num_base_bdevs_operational": 2, 00:32:53.130 "base_bdevs_list": [ 00:32:53.130 { 00:32:53.130 "name": null, 00:32:53.130 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:53.130 "is_configured": false, 00:32:53.130 "data_offset": 2048, 00:32:53.130 "data_size": 63488 00:32:53.130 }, 00:32:53.130 { 00:32:53.130 "name": "pt2", 00:32:53.130 "uuid": "00000000-0000-0000-0000-000000000002", 00:32:53.130 "is_configured": true, 00:32:53.130 "data_offset": 2048, 00:32:53.130 "data_size": 63488 00:32:53.130 }, 00:32:53.130 { 00:32:53.130 "name": null, 00:32:53.130 "uuid": "00000000-0000-0000-0000-000000000003", 00:32:53.130 "is_configured": false, 00:32:53.130 "data_offset": 2048, 00:32:53.130 "data_size": 63488 00:32:53.130 } 00:32:53.130 ] 00:32:53.130 }' 00:32:53.130 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:53.130 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:53.387 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:32:53.388 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:53.388 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:53.388 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:32:53.388 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:53.388 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:32:53.388 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:32:53.388 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:53.388 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:53.388 [2024-11-20 13:53:00.412038] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:32:53.388 [2024-11-20 13:53:00.412103] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:53.388 [2024-11-20 13:53:00.412123] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:32:53.388 [2024-11-20 13:53:00.412130] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:53.388 [2024-11-20 13:53:00.412714] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:53.388 [2024-11-20 13:53:00.412798] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:32:53.388 [2024-11-20 13:53:00.412921] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:32:53.388 [2024-11-20 13:53:00.412991] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:32:53.388 [2024-11-20 13:53:00.413126] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:32:53.388 [2024-11-20 13:53:00.413190] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:32:53.388 [2024-11-20 13:53:00.413417] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:32:53.388 [2024-11-20 13:53:00.416386] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:32:53.388 [2024-11-20 13:53:00.416456] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:32:53.388 [2024-11-20 13:53:00.416781] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:53.388 pt3 00:32:53.388 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:53.388 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:32:53.388 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:53.388 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:53.388 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:53.388 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:53.388 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:53.388 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:53.388 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:53.388 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:53.388 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:53.388 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:53.388 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:53.388 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:53.388 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:53.388 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:53.647 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:53.647 "name": "raid_bdev1", 00:32:53.647 "uuid": "9628b010-45c1-49e2-b5cd-96833ff4c37d", 00:32:53.647 "strip_size_kb": 64, 00:32:53.647 "state": "online", 00:32:53.647 "raid_level": "raid5f", 00:32:53.647 "superblock": true, 00:32:53.647 "num_base_bdevs": 3, 00:32:53.647 "num_base_bdevs_discovered": 2, 00:32:53.647 "num_base_bdevs_operational": 2, 00:32:53.647 "base_bdevs_list": [ 00:32:53.647 { 00:32:53.647 "name": null, 00:32:53.647 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:53.647 "is_configured": false, 00:32:53.647 "data_offset": 2048, 00:32:53.647 "data_size": 63488 00:32:53.647 }, 00:32:53.647 { 00:32:53.647 "name": "pt2", 00:32:53.647 "uuid": "00000000-0000-0000-0000-000000000002", 00:32:53.647 "is_configured": true, 00:32:53.647 "data_offset": 2048, 00:32:53.647 "data_size": 63488 00:32:53.647 }, 00:32:53.647 { 00:32:53.647 "name": "pt3", 00:32:53.647 "uuid": "00000000-0000-0000-0000-000000000003", 00:32:53.647 "is_configured": true, 00:32:53.647 "data_offset": 2048, 00:32:53.647 "data_size": 63488 00:32:53.647 } 00:32:53.647 ] 00:32:53.647 }' 00:32:53.647 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:53.647 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:53.647 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:32:53.647 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:32:53.647 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:53.647 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:53.906 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:53.906 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:32:53.906 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:32:53.906 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:53.906 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:53.906 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:32:53.906 [2024-11-20 13:53:00.740950] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:32:53.906 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:53.906 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 9628b010-45c1-49e2-b5cd-96833ff4c37d '!=' 9628b010-45c1-49e2-b5cd-96833ff4c37d ']' 00:32:53.906 13:53:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 78819 00:32:53.906 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 78819 ']' 00:32:53.906 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@958 -- # kill -0 78819 00:32:53.906 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # uname 00:32:53.906 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:32:53.906 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 78819 00:32:53.906 killing process with pid 78819 00:32:53.906 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:32:53.906 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:32:53.906 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 78819' 00:32:53.906 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@973 -- # kill 78819 00:32:53.906 [2024-11-20 13:53:00.795424] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:32:53.906 13:53:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@978 -- # wait 78819 00:32:53.906 [2024-11-20 13:53:00.795538] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:32:53.906 [2024-11-20 13:53:00.795599] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:32:53.906 [2024-11-20 13:53:00.795610] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:32:53.906 [2024-11-20 13:53:00.956395] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:32:54.840 13:53:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:32:54.840 00:32:54.840 real 0m5.643s 00:32:54.840 user 0m8.819s 00:32:54.840 sys 0m1.018s 00:32:54.840 13:53:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:32:54.840 13:53:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:32:54.840 ************************************ 00:32:54.840 END TEST raid5f_superblock_test 00:32:54.840 ************************************ 00:32:54.840 13:53:01 bdev_raid -- bdev/bdev_raid.sh@989 -- # '[' true = true ']' 00:32:54.840 13:53:01 bdev_raid -- bdev/bdev_raid.sh@990 -- # run_test raid5f_rebuild_test raid_rebuild_test raid5f 3 false false true 00:32:54.840 13:53:01 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:32:54.840 13:53:01 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:32:54.840 13:53:01 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:32:54.840 ************************************ 00:32:54.840 START TEST raid5f_rebuild_test 00:32:54.840 ************************************ 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 3 false false true 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=3 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=79242 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 79242 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 79242 ']' 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:54.840 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:54.840 13:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:32:54.840 [2024-11-20 13:53:01.697225] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:32:54.840 [2024-11-20 13:53:01.697534] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:32:54.840 Zero copy mechanism will not be used. 00:32:54.840 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid79242 ] 00:32:54.840 [2024-11-20 13:53:01.865764] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:55.098 [2024-11-20 13:53:01.967500] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:32:55.098 [2024-11-20 13:53:02.088825] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:32:55.098 [2024-11-20 13:53:02.088885] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:55.664 BaseBdev1_malloc 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:55.664 [2024-11-20 13:53:02.575867] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:32:55.664 [2024-11-20 13:53:02.575926] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:55.664 [2024-11-20 13:53:02.575947] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:32:55.664 [2024-11-20 13:53:02.575957] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:55.664 [2024-11-20 13:53:02.577836] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:55.664 [2024-11-20 13:53:02.577871] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:32:55.664 BaseBdev1 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:55.664 BaseBdev2_malloc 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:55.664 [2024-11-20 13:53:02.609276] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:32:55.664 [2024-11-20 13:53:02.609325] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:55.664 [2024-11-20 13:53:02.609344] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:32:55.664 [2024-11-20 13:53:02.609354] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:55.664 [2024-11-20 13:53:02.611213] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:55.664 [2024-11-20 13:53:02.611243] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:32:55.664 BaseBdev2 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:55.664 BaseBdev3_malloc 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:55.664 [2024-11-20 13:53:02.655482] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:32:55.664 [2024-11-20 13:53:02.655533] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:55.664 [2024-11-20 13:53:02.655551] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:32:55.664 [2024-11-20 13:53:02.655561] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:55.664 [2024-11-20 13:53:02.657425] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:55.664 [2024-11-20 13:53:02.657586] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:32:55.664 BaseBdev3 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:55.664 spare_malloc 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:55.664 spare_delay 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:55.664 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:55.664 [2024-11-20 13:53:02.697102] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:32:55.664 [2024-11-20 13:53:02.697147] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:55.664 [2024-11-20 13:53:02.697161] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:32:55.664 [2024-11-20 13:53:02.697170] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:55.665 [2024-11-20 13:53:02.699068] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:55.665 [2024-11-20 13:53:02.699099] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:32:55.665 spare 00:32:55.665 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:55.665 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 00:32:55.665 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:55.665 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:55.665 [2024-11-20 13:53:02.705170] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:32:55.665 [2024-11-20 13:53:02.706800] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:32:55.665 [2024-11-20 13:53:02.706856] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:32:55.665 [2024-11-20 13:53:02.706926] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:32:55.665 [2024-11-20 13:53:02.706936] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:32:55.665 [2024-11-20 13:53:02.707170] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:32:55.665 [2024-11-20 13:53:02.710236] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:32:55.665 [2024-11-20 13:53:02.710269] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:32:55.665 [2024-11-20 13:53:02.710415] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:55.665 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:55.665 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:32:55.665 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:55.665 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:55.665 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:55.665 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:55.665 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:32:55.665 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:55.665 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:55.665 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:55.665 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:55.665 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:55.665 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:55.665 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:55.665 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:55.923 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:55.923 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:55.923 "name": "raid_bdev1", 00:32:55.923 "uuid": "049e3fde-3e29-4c55-8ad4-bff8a9ac344a", 00:32:55.923 "strip_size_kb": 64, 00:32:55.923 "state": "online", 00:32:55.923 "raid_level": "raid5f", 00:32:55.923 "superblock": false, 00:32:55.923 "num_base_bdevs": 3, 00:32:55.923 "num_base_bdevs_discovered": 3, 00:32:55.923 "num_base_bdevs_operational": 3, 00:32:55.923 "base_bdevs_list": [ 00:32:55.923 { 00:32:55.923 "name": "BaseBdev1", 00:32:55.923 "uuid": "5a21ced8-c016-5ef7-ba87-6d6c306469ba", 00:32:55.923 "is_configured": true, 00:32:55.923 "data_offset": 0, 00:32:55.923 "data_size": 65536 00:32:55.923 }, 00:32:55.923 { 00:32:55.923 "name": "BaseBdev2", 00:32:55.923 "uuid": "286e34b9-2bbc-53a2-9101-093dab2a28a3", 00:32:55.923 "is_configured": true, 00:32:55.923 "data_offset": 0, 00:32:55.923 "data_size": 65536 00:32:55.923 }, 00:32:55.923 { 00:32:55.923 "name": "BaseBdev3", 00:32:55.923 "uuid": "90199e3a-b11d-5b23-bdb8-1fc0fba1cef2", 00:32:55.923 "is_configured": true, 00:32:55.923 "data_offset": 0, 00:32:55.923 "data_size": 65536 00:32:55.923 } 00:32:55.923 ] 00:32:55.923 }' 00:32:55.923 13:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:55.923 13:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:56.181 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:32:56.181 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:32:56.181 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:56.181 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:56.181 [2024-11-20 13:53:03.042696] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:32:56.181 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:56.181 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=131072 00:32:56.181 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:56.181 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:32:56.181 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:56.181 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:56.181 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:56.181 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:32:56.181 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:32:56.181 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:32:56.181 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:32:56.181 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:32:56.181 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:32:56.181 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:32:56.181 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:32:56.181 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:32:56.181 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:32:56.181 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:32:56.181 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:32:56.181 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:32:56.181 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:32:56.439 [2024-11-20 13:53:03.286598] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:32:56.439 /dev/nbd0 00:32:56.439 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:32:56.439 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:32:56.439 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:32:56.439 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:32:56.439 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:32:56.439 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:32:56.439 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:32:56.439 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:32:56.439 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:32:56.439 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:32:56.439 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:32:56.439 1+0 records in 00:32:56.439 1+0 records out 00:32:56.439 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000304158 s, 13.5 MB/s 00:32:56.439 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:32:56.439 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:32:56.439 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:32:56.439 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:32:56.440 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:32:56.440 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:32:56.440 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:32:56.440 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:32:56.440 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@630 -- # write_unit_size=256 00:32:56.440 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@631 -- # echo 128 00:32:56.440 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=131072 count=512 oflag=direct 00:32:56.697 512+0 records in 00:32:56.697 512+0 records out 00:32:56.697 67108864 bytes (67 MB, 64 MiB) copied, 0.342552 s, 196 MB/s 00:32:56.697 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:32:56.697 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:32:56.697 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:32:56.697 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:32:56.697 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:32:56.697 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:32:56.697 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:32:56.955 [2024-11-20 13:53:03.900677] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:56.955 [2024-11-20 13:53:03.908747] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:56.955 "name": "raid_bdev1", 00:32:56.955 "uuid": "049e3fde-3e29-4c55-8ad4-bff8a9ac344a", 00:32:56.955 "strip_size_kb": 64, 00:32:56.955 "state": "online", 00:32:56.955 "raid_level": "raid5f", 00:32:56.955 "superblock": false, 00:32:56.955 "num_base_bdevs": 3, 00:32:56.955 "num_base_bdevs_discovered": 2, 00:32:56.955 "num_base_bdevs_operational": 2, 00:32:56.955 "base_bdevs_list": [ 00:32:56.955 { 00:32:56.955 "name": null, 00:32:56.955 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:56.955 "is_configured": false, 00:32:56.955 "data_offset": 0, 00:32:56.955 "data_size": 65536 00:32:56.955 }, 00:32:56.955 { 00:32:56.955 "name": "BaseBdev2", 00:32:56.955 "uuid": "286e34b9-2bbc-53a2-9101-093dab2a28a3", 00:32:56.955 "is_configured": true, 00:32:56.955 "data_offset": 0, 00:32:56.955 "data_size": 65536 00:32:56.955 }, 00:32:56.955 { 00:32:56.955 "name": "BaseBdev3", 00:32:56.955 "uuid": "90199e3a-b11d-5b23-bdb8-1fc0fba1cef2", 00:32:56.955 "is_configured": true, 00:32:56.955 "data_offset": 0, 00:32:56.955 "data_size": 65536 00:32:56.955 } 00:32:56.955 ] 00:32:56.955 }' 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:56.955 13:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:57.212 13:53:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:32:57.212 13:53:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:57.212 13:53:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:57.212 [2024-11-20 13:53:04.196816] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:32:57.212 [2024-11-20 13:53:04.206200] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b680 00:32:57.212 13:53:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:57.212 13:53:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:32:57.212 [2024-11-20 13:53:04.210821] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:32:58.626 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:58.626 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:58.626 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:58.626 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:58.626 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:58.626 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:58.626 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:58.626 13:53:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:58.626 13:53:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:58.626 13:53:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:58.626 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:58.626 "name": "raid_bdev1", 00:32:58.626 "uuid": "049e3fde-3e29-4c55-8ad4-bff8a9ac344a", 00:32:58.626 "strip_size_kb": 64, 00:32:58.626 "state": "online", 00:32:58.626 "raid_level": "raid5f", 00:32:58.626 "superblock": false, 00:32:58.626 "num_base_bdevs": 3, 00:32:58.626 "num_base_bdevs_discovered": 3, 00:32:58.626 "num_base_bdevs_operational": 3, 00:32:58.626 "process": { 00:32:58.626 "type": "rebuild", 00:32:58.626 "target": "spare", 00:32:58.626 "progress": { 00:32:58.626 "blocks": 18432, 00:32:58.626 "percent": 14 00:32:58.626 } 00:32:58.626 }, 00:32:58.626 "base_bdevs_list": [ 00:32:58.626 { 00:32:58.626 "name": "spare", 00:32:58.626 "uuid": "e4027f50-653b-5ea5-9917-8fbef6e67701", 00:32:58.626 "is_configured": true, 00:32:58.626 "data_offset": 0, 00:32:58.626 "data_size": 65536 00:32:58.626 }, 00:32:58.626 { 00:32:58.626 "name": "BaseBdev2", 00:32:58.626 "uuid": "286e34b9-2bbc-53a2-9101-093dab2a28a3", 00:32:58.626 "is_configured": true, 00:32:58.626 "data_offset": 0, 00:32:58.626 "data_size": 65536 00:32:58.626 }, 00:32:58.626 { 00:32:58.626 "name": "BaseBdev3", 00:32:58.626 "uuid": "90199e3a-b11d-5b23-bdb8-1fc0fba1cef2", 00:32:58.626 "is_configured": true, 00:32:58.626 "data_offset": 0, 00:32:58.626 "data_size": 65536 00:32:58.626 } 00:32:58.626 ] 00:32:58.626 }' 00:32:58.626 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:58.626 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:58.626 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:58.626 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:58.626 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:32:58.626 13:53:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:58.626 13:53:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:58.626 [2024-11-20 13:53:05.308557] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:32:58.626 [2024-11-20 13:53:05.321786] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:32:58.626 [2024-11-20 13:53:05.321967] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:58.626 [2024-11-20 13:53:05.322003] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:32:58.626 [2024-11-20 13:53:05.322013] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:58.627 "name": "raid_bdev1", 00:32:58.627 "uuid": "049e3fde-3e29-4c55-8ad4-bff8a9ac344a", 00:32:58.627 "strip_size_kb": 64, 00:32:58.627 "state": "online", 00:32:58.627 "raid_level": "raid5f", 00:32:58.627 "superblock": false, 00:32:58.627 "num_base_bdevs": 3, 00:32:58.627 "num_base_bdevs_discovered": 2, 00:32:58.627 "num_base_bdevs_operational": 2, 00:32:58.627 "base_bdevs_list": [ 00:32:58.627 { 00:32:58.627 "name": null, 00:32:58.627 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:58.627 "is_configured": false, 00:32:58.627 "data_offset": 0, 00:32:58.627 "data_size": 65536 00:32:58.627 }, 00:32:58.627 { 00:32:58.627 "name": "BaseBdev2", 00:32:58.627 "uuid": "286e34b9-2bbc-53a2-9101-093dab2a28a3", 00:32:58.627 "is_configured": true, 00:32:58.627 "data_offset": 0, 00:32:58.627 "data_size": 65536 00:32:58.627 }, 00:32:58.627 { 00:32:58.627 "name": "BaseBdev3", 00:32:58.627 "uuid": "90199e3a-b11d-5b23-bdb8-1fc0fba1cef2", 00:32:58.627 "is_configured": true, 00:32:58.627 "data_offset": 0, 00:32:58.627 "data_size": 65536 00:32:58.627 } 00:32:58.627 ] 00:32:58.627 }' 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:58.627 13:53:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:58.913 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:58.913 "name": "raid_bdev1", 00:32:58.913 "uuid": "049e3fde-3e29-4c55-8ad4-bff8a9ac344a", 00:32:58.913 "strip_size_kb": 64, 00:32:58.913 "state": "online", 00:32:58.913 "raid_level": "raid5f", 00:32:58.913 "superblock": false, 00:32:58.913 "num_base_bdevs": 3, 00:32:58.913 "num_base_bdevs_discovered": 2, 00:32:58.913 "num_base_bdevs_operational": 2, 00:32:58.913 "base_bdevs_list": [ 00:32:58.913 { 00:32:58.913 "name": null, 00:32:58.913 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:58.913 "is_configured": false, 00:32:58.913 "data_offset": 0, 00:32:58.913 "data_size": 65536 00:32:58.913 }, 00:32:58.913 { 00:32:58.913 "name": "BaseBdev2", 00:32:58.913 "uuid": "286e34b9-2bbc-53a2-9101-093dab2a28a3", 00:32:58.913 "is_configured": true, 00:32:58.913 "data_offset": 0, 00:32:58.913 "data_size": 65536 00:32:58.913 }, 00:32:58.913 { 00:32:58.913 "name": "BaseBdev3", 00:32:58.913 "uuid": "90199e3a-b11d-5b23-bdb8-1fc0fba1cef2", 00:32:58.913 "is_configured": true, 00:32:58.913 "data_offset": 0, 00:32:58.913 "data_size": 65536 00:32:58.913 } 00:32:58.913 ] 00:32:58.913 }' 00:32:58.913 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:58.913 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:32:58.913 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:58.913 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:32:58.913 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:32:58.913 13:53:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:58.913 13:53:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:58.913 [2024-11-20 13:53:05.737244] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:32:58.913 [2024-11-20 13:53:05.745964] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b750 00:32:58.913 13:53:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:58.913 13:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:32:58.913 [2024-11-20 13:53:05.750537] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:59.846 "name": "raid_bdev1", 00:32:59.846 "uuid": "049e3fde-3e29-4c55-8ad4-bff8a9ac344a", 00:32:59.846 "strip_size_kb": 64, 00:32:59.846 "state": "online", 00:32:59.846 "raid_level": "raid5f", 00:32:59.846 "superblock": false, 00:32:59.846 "num_base_bdevs": 3, 00:32:59.846 "num_base_bdevs_discovered": 3, 00:32:59.846 "num_base_bdevs_operational": 3, 00:32:59.846 "process": { 00:32:59.846 "type": "rebuild", 00:32:59.846 "target": "spare", 00:32:59.846 "progress": { 00:32:59.846 "blocks": 18432, 00:32:59.846 "percent": 14 00:32:59.846 } 00:32:59.846 }, 00:32:59.846 "base_bdevs_list": [ 00:32:59.846 { 00:32:59.846 "name": "spare", 00:32:59.846 "uuid": "e4027f50-653b-5ea5-9917-8fbef6e67701", 00:32:59.846 "is_configured": true, 00:32:59.846 "data_offset": 0, 00:32:59.846 "data_size": 65536 00:32:59.846 }, 00:32:59.846 { 00:32:59.846 "name": "BaseBdev2", 00:32:59.846 "uuid": "286e34b9-2bbc-53a2-9101-093dab2a28a3", 00:32:59.846 "is_configured": true, 00:32:59.846 "data_offset": 0, 00:32:59.846 "data_size": 65536 00:32:59.846 }, 00:32:59.846 { 00:32:59.846 "name": "BaseBdev3", 00:32:59.846 "uuid": "90199e3a-b11d-5b23-bdb8-1fc0fba1cef2", 00:32:59.846 "is_configured": true, 00:32:59.846 "data_offset": 0, 00:32:59.846 "data_size": 65536 00:32:59.846 } 00:32:59.846 ] 00:32:59.846 }' 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=3 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=433 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:59.846 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:59.846 "name": "raid_bdev1", 00:32:59.846 "uuid": "049e3fde-3e29-4c55-8ad4-bff8a9ac344a", 00:32:59.846 "strip_size_kb": 64, 00:32:59.846 "state": "online", 00:32:59.846 "raid_level": "raid5f", 00:32:59.846 "superblock": false, 00:32:59.846 "num_base_bdevs": 3, 00:32:59.846 "num_base_bdevs_discovered": 3, 00:32:59.846 "num_base_bdevs_operational": 3, 00:32:59.846 "process": { 00:32:59.846 "type": "rebuild", 00:32:59.846 "target": "spare", 00:32:59.846 "progress": { 00:32:59.846 "blocks": 20480, 00:32:59.846 "percent": 15 00:32:59.846 } 00:32:59.846 }, 00:32:59.846 "base_bdevs_list": [ 00:32:59.846 { 00:32:59.846 "name": "spare", 00:32:59.846 "uuid": "e4027f50-653b-5ea5-9917-8fbef6e67701", 00:32:59.846 "is_configured": true, 00:32:59.846 "data_offset": 0, 00:32:59.846 "data_size": 65536 00:32:59.846 }, 00:32:59.846 { 00:32:59.847 "name": "BaseBdev2", 00:32:59.847 "uuid": "286e34b9-2bbc-53a2-9101-093dab2a28a3", 00:32:59.847 "is_configured": true, 00:32:59.847 "data_offset": 0, 00:32:59.847 "data_size": 65536 00:32:59.847 }, 00:32:59.847 { 00:32:59.847 "name": "BaseBdev3", 00:32:59.847 "uuid": "90199e3a-b11d-5b23-bdb8-1fc0fba1cef2", 00:32:59.847 "is_configured": true, 00:32:59.847 "data_offset": 0, 00:32:59.847 "data_size": 65536 00:32:59.847 } 00:32:59.847 ] 00:32:59.847 }' 00:32:59.847 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:00.104 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:33:00.104 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:00.104 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:33:00.104 13:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:33:01.037 13:53:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:33:01.037 13:53:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:01.037 13:53:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:01.038 13:53:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:01.038 13:53:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:01.038 13:53:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:01.038 13:53:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:01.038 13:53:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:01.038 13:53:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:01.038 13:53:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:01.038 13:53:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:01.038 13:53:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:01.038 "name": "raid_bdev1", 00:33:01.038 "uuid": "049e3fde-3e29-4c55-8ad4-bff8a9ac344a", 00:33:01.038 "strip_size_kb": 64, 00:33:01.038 "state": "online", 00:33:01.038 "raid_level": "raid5f", 00:33:01.038 "superblock": false, 00:33:01.038 "num_base_bdevs": 3, 00:33:01.038 "num_base_bdevs_discovered": 3, 00:33:01.038 "num_base_bdevs_operational": 3, 00:33:01.038 "process": { 00:33:01.038 "type": "rebuild", 00:33:01.038 "target": "spare", 00:33:01.038 "progress": { 00:33:01.038 "blocks": 43008, 00:33:01.038 "percent": 32 00:33:01.038 } 00:33:01.038 }, 00:33:01.038 "base_bdevs_list": [ 00:33:01.038 { 00:33:01.038 "name": "spare", 00:33:01.038 "uuid": "e4027f50-653b-5ea5-9917-8fbef6e67701", 00:33:01.038 "is_configured": true, 00:33:01.038 "data_offset": 0, 00:33:01.038 "data_size": 65536 00:33:01.038 }, 00:33:01.038 { 00:33:01.038 "name": "BaseBdev2", 00:33:01.038 "uuid": "286e34b9-2bbc-53a2-9101-093dab2a28a3", 00:33:01.038 "is_configured": true, 00:33:01.038 "data_offset": 0, 00:33:01.038 "data_size": 65536 00:33:01.038 }, 00:33:01.038 { 00:33:01.038 "name": "BaseBdev3", 00:33:01.038 "uuid": "90199e3a-b11d-5b23-bdb8-1fc0fba1cef2", 00:33:01.038 "is_configured": true, 00:33:01.038 "data_offset": 0, 00:33:01.038 "data_size": 65536 00:33:01.038 } 00:33:01.038 ] 00:33:01.038 }' 00:33:01.038 13:53:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:01.038 13:53:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:33:01.038 13:53:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:01.038 13:53:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:33:01.038 13:53:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:33:02.409 13:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:33:02.409 13:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:02.409 13:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:02.409 13:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:02.409 13:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:02.409 13:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:02.410 13:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:02.410 13:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:02.410 13:53:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:02.410 13:53:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:02.410 13:53:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:02.410 13:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:02.410 "name": "raid_bdev1", 00:33:02.410 "uuid": "049e3fde-3e29-4c55-8ad4-bff8a9ac344a", 00:33:02.410 "strip_size_kb": 64, 00:33:02.410 "state": "online", 00:33:02.410 "raid_level": "raid5f", 00:33:02.410 "superblock": false, 00:33:02.410 "num_base_bdevs": 3, 00:33:02.410 "num_base_bdevs_discovered": 3, 00:33:02.410 "num_base_bdevs_operational": 3, 00:33:02.410 "process": { 00:33:02.410 "type": "rebuild", 00:33:02.410 "target": "spare", 00:33:02.410 "progress": { 00:33:02.410 "blocks": 65536, 00:33:02.410 "percent": 50 00:33:02.410 } 00:33:02.410 }, 00:33:02.410 "base_bdevs_list": [ 00:33:02.410 { 00:33:02.410 "name": "spare", 00:33:02.410 "uuid": "e4027f50-653b-5ea5-9917-8fbef6e67701", 00:33:02.410 "is_configured": true, 00:33:02.410 "data_offset": 0, 00:33:02.410 "data_size": 65536 00:33:02.410 }, 00:33:02.410 { 00:33:02.410 "name": "BaseBdev2", 00:33:02.410 "uuid": "286e34b9-2bbc-53a2-9101-093dab2a28a3", 00:33:02.410 "is_configured": true, 00:33:02.410 "data_offset": 0, 00:33:02.410 "data_size": 65536 00:33:02.410 }, 00:33:02.410 { 00:33:02.410 "name": "BaseBdev3", 00:33:02.410 "uuid": "90199e3a-b11d-5b23-bdb8-1fc0fba1cef2", 00:33:02.410 "is_configured": true, 00:33:02.410 "data_offset": 0, 00:33:02.410 "data_size": 65536 00:33:02.410 } 00:33:02.410 ] 00:33:02.410 }' 00:33:02.410 13:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:02.410 13:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:33:02.410 13:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:02.410 13:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:33:02.410 13:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:33:03.343 13:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:33:03.343 13:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:03.343 13:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:03.343 13:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:03.343 13:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:03.344 13:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:03.344 13:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:03.344 13:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:03.344 13:53:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:03.344 13:53:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:03.344 13:53:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:03.344 13:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:03.344 "name": "raid_bdev1", 00:33:03.344 "uuid": "049e3fde-3e29-4c55-8ad4-bff8a9ac344a", 00:33:03.344 "strip_size_kb": 64, 00:33:03.344 "state": "online", 00:33:03.344 "raid_level": "raid5f", 00:33:03.344 "superblock": false, 00:33:03.344 "num_base_bdevs": 3, 00:33:03.344 "num_base_bdevs_discovered": 3, 00:33:03.344 "num_base_bdevs_operational": 3, 00:33:03.344 "process": { 00:33:03.344 "type": "rebuild", 00:33:03.344 "target": "spare", 00:33:03.344 "progress": { 00:33:03.344 "blocks": 88064, 00:33:03.344 "percent": 67 00:33:03.344 } 00:33:03.344 }, 00:33:03.344 "base_bdevs_list": [ 00:33:03.344 { 00:33:03.344 "name": "spare", 00:33:03.344 "uuid": "e4027f50-653b-5ea5-9917-8fbef6e67701", 00:33:03.344 "is_configured": true, 00:33:03.344 "data_offset": 0, 00:33:03.344 "data_size": 65536 00:33:03.344 }, 00:33:03.344 { 00:33:03.344 "name": "BaseBdev2", 00:33:03.344 "uuid": "286e34b9-2bbc-53a2-9101-093dab2a28a3", 00:33:03.344 "is_configured": true, 00:33:03.344 "data_offset": 0, 00:33:03.344 "data_size": 65536 00:33:03.344 }, 00:33:03.344 { 00:33:03.344 "name": "BaseBdev3", 00:33:03.344 "uuid": "90199e3a-b11d-5b23-bdb8-1fc0fba1cef2", 00:33:03.344 "is_configured": true, 00:33:03.344 "data_offset": 0, 00:33:03.344 "data_size": 65536 00:33:03.344 } 00:33:03.344 ] 00:33:03.344 }' 00:33:03.344 13:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:03.344 13:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:33:03.344 13:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:03.344 13:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:33:03.344 13:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:33:04.277 13:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:33:04.277 13:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:04.277 13:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:04.277 13:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:04.277 13:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:04.277 13:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:04.277 13:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:04.277 13:53:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:04.277 13:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:04.277 13:53:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:04.277 13:53:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:04.277 13:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:04.277 "name": "raid_bdev1", 00:33:04.277 "uuid": "049e3fde-3e29-4c55-8ad4-bff8a9ac344a", 00:33:04.277 "strip_size_kb": 64, 00:33:04.277 "state": "online", 00:33:04.277 "raid_level": "raid5f", 00:33:04.277 "superblock": false, 00:33:04.277 "num_base_bdevs": 3, 00:33:04.277 "num_base_bdevs_discovered": 3, 00:33:04.277 "num_base_bdevs_operational": 3, 00:33:04.277 "process": { 00:33:04.277 "type": "rebuild", 00:33:04.277 "target": "spare", 00:33:04.277 "progress": { 00:33:04.277 "blocks": 110592, 00:33:04.277 "percent": 84 00:33:04.277 } 00:33:04.277 }, 00:33:04.277 "base_bdevs_list": [ 00:33:04.277 { 00:33:04.277 "name": "spare", 00:33:04.277 "uuid": "e4027f50-653b-5ea5-9917-8fbef6e67701", 00:33:04.277 "is_configured": true, 00:33:04.277 "data_offset": 0, 00:33:04.277 "data_size": 65536 00:33:04.277 }, 00:33:04.277 { 00:33:04.277 "name": "BaseBdev2", 00:33:04.277 "uuid": "286e34b9-2bbc-53a2-9101-093dab2a28a3", 00:33:04.277 "is_configured": true, 00:33:04.277 "data_offset": 0, 00:33:04.277 "data_size": 65536 00:33:04.277 }, 00:33:04.277 { 00:33:04.277 "name": "BaseBdev3", 00:33:04.277 "uuid": "90199e3a-b11d-5b23-bdb8-1fc0fba1cef2", 00:33:04.277 "is_configured": true, 00:33:04.277 "data_offset": 0, 00:33:04.277 "data_size": 65536 00:33:04.277 } 00:33:04.277 ] 00:33:04.277 }' 00:33:04.278 13:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:04.278 13:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:33:04.278 13:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:04.535 13:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:33:04.535 13:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:33:05.469 [2024-11-20 13:53:12.212425] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:33:05.469 [2024-11-20 13:53:12.212522] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:33:05.469 [2024-11-20 13:53:12.212579] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:05.469 "name": "raid_bdev1", 00:33:05.469 "uuid": "049e3fde-3e29-4c55-8ad4-bff8a9ac344a", 00:33:05.469 "strip_size_kb": 64, 00:33:05.469 "state": "online", 00:33:05.469 "raid_level": "raid5f", 00:33:05.469 "superblock": false, 00:33:05.469 "num_base_bdevs": 3, 00:33:05.469 "num_base_bdevs_discovered": 3, 00:33:05.469 "num_base_bdevs_operational": 3, 00:33:05.469 "base_bdevs_list": [ 00:33:05.469 { 00:33:05.469 "name": "spare", 00:33:05.469 "uuid": "e4027f50-653b-5ea5-9917-8fbef6e67701", 00:33:05.469 "is_configured": true, 00:33:05.469 "data_offset": 0, 00:33:05.469 "data_size": 65536 00:33:05.469 }, 00:33:05.469 { 00:33:05.469 "name": "BaseBdev2", 00:33:05.469 "uuid": "286e34b9-2bbc-53a2-9101-093dab2a28a3", 00:33:05.469 "is_configured": true, 00:33:05.469 "data_offset": 0, 00:33:05.469 "data_size": 65536 00:33:05.469 }, 00:33:05.469 { 00:33:05.469 "name": "BaseBdev3", 00:33:05.469 "uuid": "90199e3a-b11d-5b23-bdb8-1fc0fba1cef2", 00:33:05.469 "is_configured": true, 00:33:05.469 "data_offset": 0, 00:33:05.469 "data_size": 65536 00:33:05.469 } 00:33:05.469 ] 00:33:05.469 }' 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:05.469 "name": "raid_bdev1", 00:33:05.469 "uuid": "049e3fde-3e29-4c55-8ad4-bff8a9ac344a", 00:33:05.469 "strip_size_kb": 64, 00:33:05.469 "state": "online", 00:33:05.469 "raid_level": "raid5f", 00:33:05.469 "superblock": false, 00:33:05.469 "num_base_bdevs": 3, 00:33:05.469 "num_base_bdevs_discovered": 3, 00:33:05.469 "num_base_bdevs_operational": 3, 00:33:05.469 "base_bdevs_list": [ 00:33:05.469 { 00:33:05.469 "name": "spare", 00:33:05.469 "uuid": "e4027f50-653b-5ea5-9917-8fbef6e67701", 00:33:05.469 "is_configured": true, 00:33:05.469 "data_offset": 0, 00:33:05.469 "data_size": 65536 00:33:05.469 }, 00:33:05.469 { 00:33:05.469 "name": "BaseBdev2", 00:33:05.469 "uuid": "286e34b9-2bbc-53a2-9101-093dab2a28a3", 00:33:05.469 "is_configured": true, 00:33:05.469 "data_offset": 0, 00:33:05.469 "data_size": 65536 00:33:05.469 }, 00:33:05.469 { 00:33:05.469 "name": "BaseBdev3", 00:33:05.469 "uuid": "90199e3a-b11d-5b23-bdb8-1fc0fba1cef2", 00:33:05.469 "is_configured": true, 00:33:05.469 "data_offset": 0, 00:33:05.469 "data_size": 65536 00:33:05.469 } 00:33:05.469 ] 00:33:05.469 }' 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:33:05.469 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:05.728 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:33:05.728 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:33:05.728 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:05.728 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:05.728 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:05.728 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:05.728 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:33:05.728 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:05.728 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:05.728 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:05.728 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:05.728 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:05.728 13:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:05.728 13:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:05.728 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:05.728 13:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:05.728 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:05.728 "name": "raid_bdev1", 00:33:05.728 "uuid": "049e3fde-3e29-4c55-8ad4-bff8a9ac344a", 00:33:05.728 "strip_size_kb": 64, 00:33:05.728 "state": "online", 00:33:05.728 "raid_level": "raid5f", 00:33:05.728 "superblock": false, 00:33:05.728 "num_base_bdevs": 3, 00:33:05.728 "num_base_bdevs_discovered": 3, 00:33:05.728 "num_base_bdevs_operational": 3, 00:33:05.728 "base_bdevs_list": [ 00:33:05.728 { 00:33:05.728 "name": "spare", 00:33:05.728 "uuid": "e4027f50-653b-5ea5-9917-8fbef6e67701", 00:33:05.728 "is_configured": true, 00:33:05.728 "data_offset": 0, 00:33:05.728 "data_size": 65536 00:33:05.728 }, 00:33:05.728 { 00:33:05.728 "name": "BaseBdev2", 00:33:05.728 "uuid": "286e34b9-2bbc-53a2-9101-093dab2a28a3", 00:33:05.728 "is_configured": true, 00:33:05.728 "data_offset": 0, 00:33:05.728 "data_size": 65536 00:33:05.728 }, 00:33:05.728 { 00:33:05.728 "name": "BaseBdev3", 00:33:05.728 "uuid": "90199e3a-b11d-5b23-bdb8-1fc0fba1cef2", 00:33:05.728 "is_configured": true, 00:33:05.728 "data_offset": 0, 00:33:05.728 "data_size": 65536 00:33:05.728 } 00:33:05.728 ] 00:33:05.728 }' 00:33:05.728 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:05.728 13:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:05.986 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:33:05.986 13:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:05.986 13:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:05.986 [2024-11-20 13:53:12.876221] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:33:05.986 [2024-11-20 13:53:12.876256] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:33:05.986 [2024-11-20 13:53:12.876341] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:33:05.986 [2024-11-20 13:53:12.876424] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:33:05.986 [2024-11-20 13:53:12.876438] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:33:05.986 13:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:05.986 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:05.986 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:33:05.986 13:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:05.986 13:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:05.986 13:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:05.986 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:33:05.986 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:33:05.986 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:33:05.986 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:33:05.986 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:33:05.986 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:33:05.986 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:33:05.986 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:33:05.986 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:33:05.986 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:33:05.986 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:33:05.986 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:33:05.986 13:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:33:06.244 /dev/nbd0 00:33:06.244 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:33:06.244 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:33:06.244 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:33:06.244 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:33:06.244 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:33:06.244 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:33:06.244 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:33:06.244 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:33:06.244 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:33:06.244 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:33:06.244 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:33:06.244 1+0 records in 00:33:06.244 1+0 records out 00:33:06.244 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000205424 s, 19.9 MB/s 00:33:06.244 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:33:06.245 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:33:06.245 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:33:06.245 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:33:06.245 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:33:06.245 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:33:06.245 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:33:06.245 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:33:06.503 /dev/nbd1 00:33:06.503 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:33:06.503 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:33:06.503 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:33:06.503 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:33:06.503 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:33:06.503 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:33:06.503 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:33:06.503 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:33:06.503 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:33:06.503 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:33:06.503 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:33:06.503 1+0 records in 00:33:06.503 1+0 records out 00:33:06.503 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000301097 s, 13.6 MB/s 00:33:06.503 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:33:06.503 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:33:06.503 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:33:06.503 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:33:06.503 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:33:06.503 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:33:06.503 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:33:06.503 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:33:06.503 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:33:06.503 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:33:06.503 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:33:06.503 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:33:06.503 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:33:06.503 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:33:06.503 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:33:06.761 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:33:06.761 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:33:06.761 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:33:06.761 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:33:06.761 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:33:06.761 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:33:06.761 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:33:06.761 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:33:06.761 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:33:06.761 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:33:07.018 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:33:07.019 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:33:07.019 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:33:07.019 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:33:07.019 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:33:07.019 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:33:07.019 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:33:07.019 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:33:07.019 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:33:07.019 13:53:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 79242 00:33:07.019 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 79242 ']' 00:33:07.019 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 79242 00:33:07.019 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:33:07.019 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:33:07.019 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 79242 00:33:07.019 killing process with pid 79242 00:33:07.019 Received shutdown signal, test time was about 60.000000 seconds 00:33:07.019 00:33:07.019 Latency(us) 00:33:07.019 [2024-11-20T13:53:14.078Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:33:07.019 [2024-11-20T13:53:14.078Z] =================================================================================================================== 00:33:07.019 [2024-11-20T13:53:14.078Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:33:07.019 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:33:07.019 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:33:07.019 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 79242' 00:33:07.019 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@973 -- # kill 79242 00:33:07.019 13:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@978 -- # wait 79242 00:33:07.019 [2024-11-20 13:53:13.936316] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:33:07.277 [2024-11-20 13:53:14.142702] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:33:07.843 13:53:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:33:07.843 00:33:07.843 real 0m13.133s 00:33:07.843 user 0m15.757s 00:33:07.843 sys 0m1.581s 00:33:07.843 13:53:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:33:07.843 ************************************ 00:33:07.843 END TEST raid5f_rebuild_test 00:33:07.843 ************************************ 00:33:07.843 13:53:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:07.843 13:53:14 bdev_raid -- bdev/bdev_raid.sh@991 -- # run_test raid5f_rebuild_test_sb raid_rebuild_test raid5f 3 true false true 00:33:07.843 13:53:14 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:33:07.843 13:53:14 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:33:07.843 13:53:14 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:33:07.843 ************************************ 00:33:07.843 START TEST raid5f_rebuild_test_sb 00:33:07.843 ************************************ 00:33:07.843 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 3 true false true 00:33:07.843 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:33:07.843 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=3 00:33:07.843 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:33:07.843 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:33:07.843 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:33:07.843 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:33:07.843 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:33:07.843 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:33:07.843 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:33:07.843 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:33:07.843 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:33:07.843 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:33:07.843 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:33:07.843 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:33:07.843 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:33:07.844 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:33:07.844 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:33:07.844 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:33:07.844 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:33:07.844 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:33:07.844 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:33:07.844 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:33:07.844 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:33:07.844 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:33:07.844 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:33:07.844 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:33:07.844 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:33:07.844 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:33:07.844 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:33:07.844 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=79660 00:33:07.844 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 79660 00:33:07.844 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:33:07.844 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 79660 ']' 00:33:07.844 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:33:07.844 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:33:07.844 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:33:07.844 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:33:07.844 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:33:07.844 13:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:07.844 [2024-11-20 13:53:14.878860] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:33:07.844 I/O size of 3145728 is greater than zero copy threshold (65536). 00:33:07.844 Zero copy mechanism will not be used. 00:33:07.844 [2024-11-20 13:53:14.879139] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid79660 ] 00:33:08.102 [2024-11-20 13:53:15.035112] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:33:08.103 [2024-11-20 13:53:15.137070] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:33:08.361 [2024-11-20 13:53:15.259642] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:33:08.361 [2024-11-20 13:53:15.259686] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:08.927 BaseBdev1_malloc 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:08.927 [2024-11-20 13:53:15.743880] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:33:08.927 [2024-11-20 13:53:15.744108] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:08.927 [2024-11-20 13:53:15.744133] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:33:08.927 [2024-11-20 13:53:15.744144] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:08.927 [2024-11-20 13:53:15.745996] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:08.927 [2024-11-20 13:53:15.746026] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:33:08.927 BaseBdev1 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:08.927 BaseBdev2_malloc 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:08.927 [2024-11-20 13:53:15.777199] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:33:08.927 [2024-11-20 13:53:15.777254] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:08.927 [2024-11-20 13:53:15.777273] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:33:08.927 [2024-11-20 13:53:15.777282] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:08.927 [2024-11-20 13:53:15.779146] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:08.927 [2024-11-20 13:53:15.779295] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:33:08.927 BaseBdev2 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:08.927 BaseBdev3_malloc 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:08.927 [2024-11-20 13:53:15.835650] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:33:08.927 [2024-11-20 13:53:15.835711] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:08.927 [2024-11-20 13:53:15.835733] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:33:08.927 [2024-11-20 13:53:15.835743] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:08.927 [2024-11-20 13:53:15.837668] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:08.927 [2024-11-20 13:53:15.837702] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:33:08.927 BaseBdev3 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:08.927 spare_malloc 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:08.927 spare_delay 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:33:08.927 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:08.928 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:08.928 [2024-11-20 13:53:15.881491] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:33:08.928 [2024-11-20 13:53:15.881548] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:08.928 [2024-11-20 13:53:15.881564] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:33:08.928 [2024-11-20 13:53:15.881575] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:08.928 [2024-11-20 13:53:15.883494] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:08.928 [2024-11-20 13:53:15.883638] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:33:08.928 spare 00:33:08.928 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:08.928 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 00:33:08.928 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:08.928 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:08.928 [2024-11-20 13:53:15.889558] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:33:08.928 [2024-11-20 13:53:15.891286] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:33:08.928 [2024-11-20 13:53:15.891395] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:33:08.928 [2024-11-20 13:53:15.891565] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:33:08.928 [2024-11-20 13:53:15.891589] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:33:08.928 [2024-11-20 13:53:15.891855] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:33:08.928 [2024-11-20 13:53:15.895012] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:33:08.928 [2024-11-20 13:53:15.895092] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:33:08.928 [2024-11-20 13:53:15.895299] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:08.928 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:08.928 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:33:08.928 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:08.928 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:08.928 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:08.928 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:08.928 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:33:08.928 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:08.928 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:08.928 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:08.928 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:08.928 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:08.928 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:08.928 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:08.928 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:08.928 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:08.928 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:08.928 "name": "raid_bdev1", 00:33:08.928 "uuid": "04a83dac-3571-4037-8e53-5786897ec384", 00:33:08.928 "strip_size_kb": 64, 00:33:08.928 "state": "online", 00:33:08.928 "raid_level": "raid5f", 00:33:08.928 "superblock": true, 00:33:08.928 "num_base_bdevs": 3, 00:33:08.928 "num_base_bdevs_discovered": 3, 00:33:08.928 "num_base_bdevs_operational": 3, 00:33:08.928 "base_bdevs_list": [ 00:33:08.928 { 00:33:08.928 "name": "BaseBdev1", 00:33:08.928 "uuid": "c1c0962b-e301-5367-a2ee-02ca102c1ad1", 00:33:08.928 "is_configured": true, 00:33:08.928 "data_offset": 2048, 00:33:08.928 "data_size": 63488 00:33:08.928 }, 00:33:08.928 { 00:33:08.928 "name": "BaseBdev2", 00:33:08.928 "uuid": "45032530-0c80-5314-b7cd-d3e3d498c9e8", 00:33:08.928 "is_configured": true, 00:33:08.928 "data_offset": 2048, 00:33:08.928 "data_size": 63488 00:33:08.928 }, 00:33:08.928 { 00:33:08.928 "name": "BaseBdev3", 00:33:08.928 "uuid": "cae27a91-cd22-5c5e-8ee4-b6d0cfd7d514", 00:33:08.928 "is_configured": true, 00:33:08.928 "data_offset": 2048, 00:33:08.928 "data_size": 63488 00:33:08.928 } 00:33:08.928 ] 00:33:08.928 }' 00:33:08.928 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:08.928 13:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:09.186 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:33:09.187 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:33:09.187 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:09.187 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:09.187 [2024-11-20 13:53:16.214445] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:33:09.187 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:09.187 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=126976 00:33:09.187 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:33:09.187 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:09.187 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:09.187 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:09.444 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:09.444 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:33:09.444 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:33:09.444 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:33:09.445 [2024-11-20 13:53:16.446353] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:33:09.445 /dev/nbd0 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:33:09.445 1+0 records in 00:33:09.445 1+0 records out 00:33:09.445 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000414798 s, 9.9 MB/s 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:33:09.445 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:33:09.702 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:33:09.702 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:33:09.702 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:33:09.702 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@630 -- # write_unit_size=256 00:33:09.702 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@631 -- # echo 128 00:33:09.702 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=131072 count=496 oflag=direct 00:33:09.960 496+0 records in 00:33:09.960 496+0 records out 00:33:09.960 65011712 bytes (65 MB, 62 MiB) copied, 0.36535 s, 178 MB/s 00:33:09.960 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:33:09.960 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:33:09.960 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:33:09.960 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:33:09.960 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:33:09.960 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:33:09.960 13:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:33:10.217 [2024-11-20 13:53:17.066901] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:10.217 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:33:10.217 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:33:10.217 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:33:10.217 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:33:10.217 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:33:10.217 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:33:10.217 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:33:10.217 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:33:10.217 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:33:10.217 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:10.217 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:10.217 [2024-11-20 13:53:17.099352] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:33:10.218 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:10.218 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:33:10.218 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:10.218 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:10.218 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:10.218 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:10.218 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:33:10.218 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:10.218 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:10.218 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:10.218 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:10.218 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:10.218 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:10.218 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:10.218 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:10.218 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:10.218 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:10.218 "name": "raid_bdev1", 00:33:10.218 "uuid": "04a83dac-3571-4037-8e53-5786897ec384", 00:33:10.218 "strip_size_kb": 64, 00:33:10.218 "state": "online", 00:33:10.218 "raid_level": "raid5f", 00:33:10.218 "superblock": true, 00:33:10.218 "num_base_bdevs": 3, 00:33:10.218 "num_base_bdevs_discovered": 2, 00:33:10.218 "num_base_bdevs_operational": 2, 00:33:10.218 "base_bdevs_list": [ 00:33:10.218 { 00:33:10.218 "name": null, 00:33:10.218 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:10.218 "is_configured": false, 00:33:10.218 "data_offset": 0, 00:33:10.218 "data_size": 63488 00:33:10.218 }, 00:33:10.218 { 00:33:10.218 "name": "BaseBdev2", 00:33:10.218 "uuid": "45032530-0c80-5314-b7cd-d3e3d498c9e8", 00:33:10.218 "is_configured": true, 00:33:10.218 "data_offset": 2048, 00:33:10.218 "data_size": 63488 00:33:10.218 }, 00:33:10.218 { 00:33:10.218 "name": "BaseBdev3", 00:33:10.218 "uuid": "cae27a91-cd22-5c5e-8ee4-b6d0cfd7d514", 00:33:10.218 "is_configured": true, 00:33:10.218 "data_offset": 2048, 00:33:10.218 "data_size": 63488 00:33:10.218 } 00:33:10.218 ] 00:33:10.218 }' 00:33:10.218 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:10.218 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:10.475 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:33:10.476 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:10.476 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:10.476 [2024-11-20 13:53:17.419430] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:33:10.476 [2024-11-20 13:53:17.430855] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000028f80 00:33:10.476 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:10.476 13:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:33:10.476 [2024-11-20 13:53:17.436535] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:33:11.409 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:11.409 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:11.409 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:11.409 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:11.409 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:11.409 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:11.409 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:11.409 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:11.409 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:11.409 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:11.667 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:11.667 "name": "raid_bdev1", 00:33:11.667 "uuid": "04a83dac-3571-4037-8e53-5786897ec384", 00:33:11.667 "strip_size_kb": 64, 00:33:11.667 "state": "online", 00:33:11.667 "raid_level": "raid5f", 00:33:11.667 "superblock": true, 00:33:11.667 "num_base_bdevs": 3, 00:33:11.667 "num_base_bdevs_discovered": 3, 00:33:11.667 "num_base_bdevs_operational": 3, 00:33:11.667 "process": { 00:33:11.667 "type": "rebuild", 00:33:11.667 "target": "spare", 00:33:11.667 "progress": { 00:33:11.667 "blocks": 18432, 00:33:11.667 "percent": 14 00:33:11.667 } 00:33:11.667 }, 00:33:11.667 "base_bdevs_list": [ 00:33:11.667 { 00:33:11.667 "name": "spare", 00:33:11.667 "uuid": "06678d61-a1f4-5eb1-801a-208b14fcb645", 00:33:11.667 "is_configured": true, 00:33:11.667 "data_offset": 2048, 00:33:11.667 "data_size": 63488 00:33:11.667 }, 00:33:11.667 { 00:33:11.667 "name": "BaseBdev2", 00:33:11.667 "uuid": "45032530-0c80-5314-b7cd-d3e3d498c9e8", 00:33:11.667 "is_configured": true, 00:33:11.667 "data_offset": 2048, 00:33:11.667 "data_size": 63488 00:33:11.667 }, 00:33:11.667 { 00:33:11.667 "name": "BaseBdev3", 00:33:11.667 "uuid": "cae27a91-cd22-5c5e-8ee4-b6d0cfd7d514", 00:33:11.667 "is_configured": true, 00:33:11.667 "data_offset": 2048, 00:33:11.667 "data_size": 63488 00:33:11.667 } 00:33:11.667 ] 00:33:11.667 }' 00:33:11.667 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:11.667 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:33:11.667 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:11.667 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:33:11.667 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:33:11.667 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:11.667 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:11.667 [2024-11-20 13:53:18.533663] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:33:11.667 [2024-11-20 13:53:18.548709] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:33:11.667 [2024-11-20 13:53:18.548776] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:11.667 [2024-11-20 13:53:18.548796] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:33:11.667 [2024-11-20 13:53:18.548805] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:33:11.667 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:11.667 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:33:11.667 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:11.667 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:11.667 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:11.667 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:11.667 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:33:11.667 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:11.667 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:11.667 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:11.667 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:11.667 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:11.667 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:11.667 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:11.667 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:11.667 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:11.667 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:11.667 "name": "raid_bdev1", 00:33:11.667 "uuid": "04a83dac-3571-4037-8e53-5786897ec384", 00:33:11.667 "strip_size_kb": 64, 00:33:11.667 "state": "online", 00:33:11.667 "raid_level": "raid5f", 00:33:11.667 "superblock": true, 00:33:11.667 "num_base_bdevs": 3, 00:33:11.667 "num_base_bdevs_discovered": 2, 00:33:11.667 "num_base_bdevs_operational": 2, 00:33:11.667 "base_bdevs_list": [ 00:33:11.667 { 00:33:11.667 "name": null, 00:33:11.667 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:11.667 "is_configured": false, 00:33:11.667 "data_offset": 0, 00:33:11.667 "data_size": 63488 00:33:11.668 }, 00:33:11.668 { 00:33:11.668 "name": "BaseBdev2", 00:33:11.668 "uuid": "45032530-0c80-5314-b7cd-d3e3d498c9e8", 00:33:11.668 "is_configured": true, 00:33:11.668 "data_offset": 2048, 00:33:11.668 "data_size": 63488 00:33:11.668 }, 00:33:11.668 { 00:33:11.668 "name": "BaseBdev3", 00:33:11.668 "uuid": "cae27a91-cd22-5c5e-8ee4-b6d0cfd7d514", 00:33:11.668 "is_configured": true, 00:33:11.668 "data_offset": 2048, 00:33:11.668 "data_size": 63488 00:33:11.668 } 00:33:11.668 ] 00:33:11.668 }' 00:33:11.668 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:11.668 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:11.925 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:33:11.926 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:11.926 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:33:11.926 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:33:11.926 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:11.926 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:11.926 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:11.926 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:11.926 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:11.926 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:11.926 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:11.926 "name": "raid_bdev1", 00:33:11.926 "uuid": "04a83dac-3571-4037-8e53-5786897ec384", 00:33:11.926 "strip_size_kb": 64, 00:33:11.926 "state": "online", 00:33:11.926 "raid_level": "raid5f", 00:33:11.926 "superblock": true, 00:33:11.926 "num_base_bdevs": 3, 00:33:11.926 "num_base_bdevs_discovered": 2, 00:33:11.926 "num_base_bdevs_operational": 2, 00:33:11.926 "base_bdevs_list": [ 00:33:11.926 { 00:33:11.926 "name": null, 00:33:11.926 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:11.926 "is_configured": false, 00:33:11.926 "data_offset": 0, 00:33:11.926 "data_size": 63488 00:33:11.926 }, 00:33:11.926 { 00:33:11.926 "name": "BaseBdev2", 00:33:11.926 "uuid": "45032530-0c80-5314-b7cd-d3e3d498c9e8", 00:33:11.926 "is_configured": true, 00:33:11.926 "data_offset": 2048, 00:33:11.926 "data_size": 63488 00:33:11.926 }, 00:33:11.926 { 00:33:11.926 "name": "BaseBdev3", 00:33:11.926 "uuid": "cae27a91-cd22-5c5e-8ee4-b6d0cfd7d514", 00:33:11.926 "is_configured": true, 00:33:11.926 "data_offset": 2048, 00:33:11.926 "data_size": 63488 00:33:11.926 } 00:33:11.926 ] 00:33:11.926 }' 00:33:11.926 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:12.184 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:33:12.184 13:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:12.184 13:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:33:12.184 13:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:33:12.184 13:53:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:12.184 13:53:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:12.184 [2024-11-20 13:53:19.033418] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:33:12.184 [2024-11-20 13:53:19.042132] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000029050 00:33:12.184 13:53:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:12.184 13:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:33:12.184 [2024-11-20 13:53:19.046585] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:33:13.164 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:13.164 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:13.164 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:13.164 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:13.164 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:13.164 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:13.164 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:13.164 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:13.164 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:13.164 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:13.164 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:13.164 "name": "raid_bdev1", 00:33:13.164 "uuid": "04a83dac-3571-4037-8e53-5786897ec384", 00:33:13.164 "strip_size_kb": 64, 00:33:13.164 "state": "online", 00:33:13.164 "raid_level": "raid5f", 00:33:13.164 "superblock": true, 00:33:13.164 "num_base_bdevs": 3, 00:33:13.164 "num_base_bdevs_discovered": 3, 00:33:13.164 "num_base_bdevs_operational": 3, 00:33:13.164 "process": { 00:33:13.164 "type": "rebuild", 00:33:13.164 "target": "spare", 00:33:13.164 "progress": { 00:33:13.164 "blocks": 18432, 00:33:13.164 "percent": 14 00:33:13.164 } 00:33:13.164 }, 00:33:13.164 "base_bdevs_list": [ 00:33:13.164 { 00:33:13.164 "name": "spare", 00:33:13.164 "uuid": "06678d61-a1f4-5eb1-801a-208b14fcb645", 00:33:13.164 "is_configured": true, 00:33:13.164 "data_offset": 2048, 00:33:13.164 "data_size": 63488 00:33:13.164 }, 00:33:13.164 { 00:33:13.164 "name": "BaseBdev2", 00:33:13.164 "uuid": "45032530-0c80-5314-b7cd-d3e3d498c9e8", 00:33:13.164 "is_configured": true, 00:33:13.164 "data_offset": 2048, 00:33:13.164 "data_size": 63488 00:33:13.164 }, 00:33:13.164 { 00:33:13.164 "name": "BaseBdev3", 00:33:13.164 "uuid": "cae27a91-cd22-5c5e-8ee4-b6d0cfd7d514", 00:33:13.165 "is_configured": true, 00:33:13.165 "data_offset": 2048, 00:33:13.165 "data_size": 63488 00:33:13.165 } 00:33:13.165 ] 00:33:13.165 }' 00:33:13.165 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:13.165 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:33:13.165 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:13.165 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:33:13.165 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:33:13.165 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:33:13.165 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:33:13.165 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=3 00:33:13.165 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:33:13.165 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=447 00:33:13.165 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:33:13.165 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:13.165 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:13.165 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:13.165 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:13.165 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:13.165 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:13.165 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:13.165 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:13.165 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:13.165 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:13.165 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:13.165 "name": "raid_bdev1", 00:33:13.165 "uuid": "04a83dac-3571-4037-8e53-5786897ec384", 00:33:13.165 "strip_size_kb": 64, 00:33:13.165 "state": "online", 00:33:13.165 "raid_level": "raid5f", 00:33:13.165 "superblock": true, 00:33:13.165 "num_base_bdevs": 3, 00:33:13.165 "num_base_bdevs_discovered": 3, 00:33:13.165 "num_base_bdevs_operational": 3, 00:33:13.165 "process": { 00:33:13.165 "type": "rebuild", 00:33:13.165 "target": "spare", 00:33:13.165 "progress": { 00:33:13.165 "blocks": 22528, 00:33:13.165 "percent": 17 00:33:13.165 } 00:33:13.165 }, 00:33:13.165 "base_bdevs_list": [ 00:33:13.165 { 00:33:13.165 "name": "spare", 00:33:13.165 "uuid": "06678d61-a1f4-5eb1-801a-208b14fcb645", 00:33:13.165 "is_configured": true, 00:33:13.165 "data_offset": 2048, 00:33:13.165 "data_size": 63488 00:33:13.165 }, 00:33:13.165 { 00:33:13.165 "name": "BaseBdev2", 00:33:13.165 "uuid": "45032530-0c80-5314-b7cd-d3e3d498c9e8", 00:33:13.165 "is_configured": true, 00:33:13.165 "data_offset": 2048, 00:33:13.165 "data_size": 63488 00:33:13.165 }, 00:33:13.165 { 00:33:13.165 "name": "BaseBdev3", 00:33:13.165 "uuid": "cae27a91-cd22-5c5e-8ee4-b6d0cfd7d514", 00:33:13.165 "is_configured": true, 00:33:13.165 "data_offset": 2048, 00:33:13.165 "data_size": 63488 00:33:13.165 } 00:33:13.165 ] 00:33:13.165 }' 00:33:13.165 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:13.165 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:33:13.165 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:13.424 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:33:13.424 13:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:33:14.358 13:53:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:33:14.358 13:53:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:14.358 13:53:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:14.358 13:53:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:14.358 13:53:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:14.358 13:53:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:14.358 13:53:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:14.358 13:53:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:14.358 13:53:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:14.358 13:53:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:14.358 13:53:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:14.358 13:53:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:14.358 "name": "raid_bdev1", 00:33:14.358 "uuid": "04a83dac-3571-4037-8e53-5786897ec384", 00:33:14.358 "strip_size_kb": 64, 00:33:14.358 "state": "online", 00:33:14.358 "raid_level": "raid5f", 00:33:14.358 "superblock": true, 00:33:14.358 "num_base_bdevs": 3, 00:33:14.358 "num_base_bdevs_discovered": 3, 00:33:14.358 "num_base_bdevs_operational": 3, 00:33:14.358 "process": { 00:33:14.358 "type": "rebuild", 00:33:14.358 "target": "spare", 00:33:14.358 "progress": { 00:33:14.358 "blocks": 43008, 00:33:14.358 "percent": 33 00:33:14.358 } 00:33:14.358 }, 00:33:14.358 "base_bdevs_list": [ 00:33:14.358 { 00:33:14.358 "name": "spare", 00:33:14.358 "uuid": "06678d61-a1f4-5eb1-801a-208b14fcb645", 00:33:14.358 "is_configured": true, 00:33:14.358 "data_offset": 2048, 00:33:14.358 "data_size": 63488 00:33:14.358 }, 00:33:14.358 { 00:33:14.358 "name": "BaseBdev2", 00:33:14.358 "uuid": "45032530-0c80-5314-b7cd-d3e3d498c9e8", 00:33:14.358 "is_configured": true, 00:33:14.358 "data_offset": 2048, 00:33:14.358 "data_size": 63488 00:33:14.358 }, 00:33:14.358 { 00:33:14.358 "name": "BaseBdev3", 00:33:14.358 "uuid": "cae27a91-cd22-5c5e-8ee4-b6d0cfd7d514", 00:33:14.358 "is_configured": true, 00:33:14.358 "data_offset": 2048, 00:33:14.358 "data_size": 63488 00:33:14.358 } 00:33:14.358 ] 00:33:14.358 }' 00:33:14.358 13:53:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:14.358 13:53:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:33:14.358 13:53:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:14.358 13:53:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:33:14.358 13:53:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:33:15.729 13:53:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:33:15.729 13:53:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:15.729 13:53:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:15.729 13:53:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:15.729 13:53:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:15.730 13:53:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:15.730 13:53:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:15.730 13:53:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:15.730 13:53:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:15.730 13:53:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:15.730 13:53:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:15.730 13:53:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:15.730 "name": "raid_bdev1", 00:33:15.730 "uuid": "04a83dac-3571-4037-8e53-5786897ec384", 00:33:15.730 "strip_size_kb": 64, 00:33:15.730 "state": "online", 00:33:15.730 "raid_level": "raid5f", 00:33:15.730 "superblock": true, 00:33:15.730 "num_base_bdevs": 3, 00:33:15.730 "num_base_bdevs_discovered": 3, 00:33:15.730 "num_base_bdevs_operational": 3, 00:33:15.730 "process": { 00:33:15.730 "type": "rebuild", 00:33:15.730 "target": "spare", 00:33:15.730 "progress": { 00:33:15.730 "blocks": 65536, 00:33:15.730 "percent": 51 00:33:15.730 } 00:33:15.730 }, 00:33:15.730 "base_bdevs_list": [ 00:33:15.730 { 00:33:15.730 "name": "spare", 00:33:15.730 "uuid": "06678d61-a1f4-5eb1-801a-208b14fcb645", 00:33:15.730 "is_configured": true, 00:33:15.730 "data_offset": 2048, 00:33:15.730 "data_size": 63488 00:33:15.730 }, 00:33:15.730 { 00:33:15.730 "name": "BaseBdev2", 00:33:15.730 "uuid": "45032530-0c80-5314-b7cd-d3e3d498c9e8", 00:33:15.730 "is_configured": true, 00:33:15.730 "data_offset": 2048, 00:33:15.730 "data_size": 63488 00:33:15.730 }, 00:33:15.730 { 00:33:15.730 "name": "BaseBdev3", 00:33:15.730 "uuid": "cae27a91-cd22-5c5e-8ee4-b6d0cfd7d514", 00:33:15.730 "is_configured": true, 00:33:15.730 "data_offset": 2048, 00:33:15.730 "data_size": 63488 00:33:15.730 } 00:33:15.730 ] 00:33:15.730 }' 00:33:15.730 13:53:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:15.730 13:53:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:33:15.730 13:53:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:15.730 13:53:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:33:15.730 13:53:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:33:16.677 13:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:33:16.677 13:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:16.677 13:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:16.677 13:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:16.677 13:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:16.677 13:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:16.677 13:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:16.677 13:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:16.677 13:53:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:16.677 13:53:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:16.677 13:53:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:16.677 13:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:16.677 "name": "raid_bdev1", 00:33:16.677 "uuid": "04a83dac-3571-4037-8e53-5786897ec384", 00:33:16.677 "strip_size_kb": 64, 00:33:16.677 "state": "online", 00:33:16.677 "raid_level": "raid5f", 00:33:16.677 "superblock": true, 00:33:16.677 "num_base_bdevs": 3, 00:33:16.677 "num_base_bdevs_discovered": 3, 00:33:16.677 "num_base_bdevs_operational": 3, 00:33:16.677 "process": { 00:33:16.677 "type": "rebuild", 00:33:16.677 "target": "spare", 00:33:16.677 "progress": { 00:33:16.677 "blocks": 88064, 00:33:16.677 "percent": 69 00:33:16.677 } 00:33:16.677 }, 00:33:16.677 "base_bdevs_list": [ 00:33:16.677 { 00:33:16.677 "name": "spare", 00:33:16.677 "uuid": "06678d61-a1f4-5eb1-801a-208b14fcb645", 00:33:16.678 "is_configured": true, 00:33:16.678 "data_offset": 2048, 00:33:16.678 "data_size": 63488 00:33:16.678 }, 00:33:16.678 { 00:33:16.678 "name": "BaseBdev2", 00:33:16.678 "uuid": "45032530-0c80-5314-b7cd-d3e3d498c9e8", 00:33:16.678 "is_configured": true, 00:33:16.678 "data_offset": 2048, 00:33:16.678 "data_size": 63488 00:33:16.678 }, 00:33:16.678 { 00:33:16.678 "name": "BaseBdev3", 00:33:16.678 "uuid": "cae27a91-cd22-5c5e-8ee4-b6d0cfd7d514", 00:33:16.678 "is_configured": true, 00:33:16.678 "data_offset": 2048, 00:33:16.678 "data_size": 63488 00:33:16.678 } 00:33:16.678 ] 00:33:16.678 }' 00:33:16.678 13:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:16.678 13:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:33:16.678 13:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:16.678 13:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:33:16.678 13:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:33:17.612 13:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:33:17.612 13:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:17.612 13:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:17.612 13:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:17.612 13:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:17.612 13:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:17.612 13:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:17.612 13:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:17.612 13:53:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:17.612 13:53:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:17.612 13:53:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:17.612 13:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:17.612 "name": "raid_bdev1", 00:33:17.612 "uuid": "04a83dac-3571-4037-8e53-5786897ec384", 00:33:17.612 "strip_size_kb": 64, 00:33:17.612 "state": "online", 00:33:17.612 "raid_level": "raid5f", 00:33:17.612 "superblock": true, 00:33:17.612 "num_base_bdevs": 3, 00:33:17.612 "num_base_bdevs_discovered": 3, 00:33:17.612 "num_base_bdevs_operational": 3, 00:33:17.612 "process": { 00:33:17.612 "type": "rebuild", 00:33:17.612 "target": "spare", 00:33:17.612 "progress": { 00:33:17.612 "blocks": 110592, 00:33:17.612 "percent": 87 00:33:17.612 } 00:33:17.612 }, 00:33:17.612 "base_bdevs_list": [ 00:33:17.612 { 00:33:17.612 "name": "spare", 00:33:17.612 "uuid": "06678d61-a1f4-5eb1-801a-208b14fcb645", 00:33:17.612 "is_configured": true, 00:33:17.612 "data_offset": 2048, 00:33:17.612 "data_size": 63488 00:33:17.612 }, 00:33:17.612 { 00:33:17.612 "name": "BaseBdev2", 00:33:17.612 "uuid": "45032530-0c80-5314-b7cd-d3e3d498c9e8", 00:33:17.612 "is_configured": true, 00:33:17.612 "data_offset": 2048, 00:33:17.612 "data_size": 63488 00:33:17.612 }, 00:33:17.612 { 00:33:17.612 "name": "BaseBdev3", 00:33:17.612 "uuid": "cae27a91-cd22-5c5e-8ee4-b6d0cfd7d514", 00:33:17.612 "is_configured": true, 00:33:17.612 "data_offset": 2048, 00:33:17.612 "data_size": 63488 00:33:17.612 } 00:33:17.612 ] 00:33:17.612 }' 00:33:17.612 13:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:17.612 13:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:33:17.612 13:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:17.612 13:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:33:17.612 13:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:33:18.546 [2024-11-20 13:53:25.306029] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:33:18.546 [2024-11-20 13:53:25.306134] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:33:18.546 [2024-11-20 13:53:25.306271] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:18.804 "name": "raid_bdev1", 00:33:18.804 "uuid": "04a83dac-3571-4037-8e53-5786897ec384", 00:33:18.804 "strip_size_kb": 64, 00:33:18.804 "state": "online", 00:33:18.804 "raid_level": "raid5f", 00:33:18.804 "superblock": true, 00:33:18.804 "num_base_bdevs": 3, 00:33:18.804 "num_base_bdevs_discovered": 3, 00:33:18.804 "num_base_bdevs_operational": 3, 00:33:18.804 "base_bdevs_list": [ 00:33:18.804 { 00:33:18.804 "name": "spare", 00:33:18.804 "uuid": "06678d61-a1f4-5eb1-801a-208b14fcb645", 00:33:18.804 "is_configured": true, 00:33:18.804 "data_offset": 2048, 00:33:18.804 "data_size": 63488 00:33:18.804 }, 00:33:18.804 { 00:33:18.804 "name": "BaseBdev2", 00:33:18.804 "uuid": "45032530-0c80-5314-b7cd-d3e3d498c9e8", 00:33:18.804 "is_configured": true, 00:33:18.804 "data_offset": 2048, 00:33:18.804 "data_size": 63488 00:33:18.804 }, 00:33:18.804 { 00:33:18.804 "name": "BaseBdev3", 00:33:18.804 "uuid": "cae27a91-cd22-5c5e-8ee4-b6d0cfd7d514", 00:33:18.804 "is_configured": true, 00:33:18.804 "data_offset": 2048, 00:33:18.804 "data_size": 63488 00:33:18.804 } 00:33:18.804 ] 00:33:18.804 }' 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:18.804 "name": "raid_bdev1", 00:33:18.804 "uuid": "04a83dac-3571-4037-8e53-5786897ec384", 00:33:18.804 "strip_size_kb": 64, 00:33:18.804 "state": "online", 00:33:18.804 "raid_level": "raid5f", 00:33:18.804 "superblock": true, 00:33:18.804 "num_base_bdevs": 3, 00:33:18.804 "num_base_bdevs_discovered": 3, 00:33:18.804 "num_base_bdevs_operational": 3, 00:33:18.804 "base_bdevs_list": [ 00:33:18.804 { 00:33:18.804 "name": "spare", 00:33:18.804 "uuid": "06678d61-a1f4-5eb1-801a-208b14fcb645", 00:33:18.804 "is_configured": true, 00:33:18.804 "data_offset": 2048, 00:33:18.804 "data_size": 63488 00:33:18.804 }, 00:33:18.804 { 00:33:18.804 "name": "BaseBdev2", 00:33:18.804 "uuid": "45032530-0c80-5314-b7cd-d3e3d498c9e8", 00:33:18.804 "is_configured": true, 00:33:18.804 "data_offset": 2048, 00:33:18.804 "data_size": 63488 00:33:18.804 }, 00:33:18.804 { 00:33:18.804 "name": "BaseBdev3", 00:33:18.804 "uuid": "cae27a91-cd22-5c5e-8ee4-b6d0cfd7d514", 00:33:18.804 "is_configured": true, 00:33:18.804 "data_offset": 2048, 00:33:18.804 "data_size": 63488 00:33:18.804 } 00:33:18.804 ] 00:33:18.804 }' 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:33:18.804 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:19.063 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:33:19.063 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:33:19.063 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:19.063 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:19.063 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:19.063 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:19.063 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:33:19.063 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:19.063 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:19.063 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:19.063 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:19.063 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:19.063 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:19.063 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:19.063 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:19.063 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:19.063 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:19.063 "name": "raid_bdev1", 00:33:19.063 "uuid": "04a83dac-3571-4037-8e53-5786897ec384", 00:33:19.063 "strip_size_kb": 64, 00:33:19.063 "state": "online", 00:33:19.063 "raid_level": "raid5f", 00:33:19.063 "superblock": true, 00:33:19.063 "num_base_bdevs": 3, 00:33:19.063 "num_base_bdevs_discovered": 3, 00:33:19.063 "num_base_bdevs_operational": 3, 00:33:19.063 "base_bdevs_list": [ 00:33:19.063 { 00:33:19.063 "name": "spare", 00:33:19.063 "uuid": "06678d61-a1f4-5eb1-801a-208b14fcb645", 00:33:19.063 "is_configured": true, 00:33:19.063 "data_offset": 2048, 00:33:19.063 "data_size": 63488 00:33:19.063 }, 00:33:19.063 { 00:33:19.063 "name": "BaseBdev2", 00:33:19.063 "uuid": "45032530-0c80-5314-b7cd-d3e3d498c9e8", 00:33:19.063 "is_configured": true, 00:33:19.063 "data_offset": 2048, 00:33:19.063 "data_size": 63488 00:33:19.063 }, 00:33:19.063 { 00:33:19.063 "name": "BaseBdev3", 00:33:19.063 "uuid": "cae27a91-cd22-5c5e-8ee4-b6d0cfd7d514", 00:33:19.063 "is_configured": true, 00:33:19.063 "data_offset": 2048, 00:33:19.063 "data_size": 63488 00:33:19.063 } 00:33:19.063 ] 00:33:19.063 }' 00:33:19.063 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:19.063 13:53:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:19.321 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:33:19.321 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:19.321 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:19.321 [2024-11-20 13:53:26.189328] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:33:19.321 [2024-11-20 13:53:26.189475] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:33:19.321 [2024-11-20 13:53:26.189614] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:33:19.321 [2024-11-20 13:53:26.189747] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:33:19.321 [2024-11-20 13:53:26.189817] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:33:19.321 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:19.321 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:19.321 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:19.321 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:19.321 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:33:19.321 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:19.321 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:33:19.321 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:33:19.321 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:33:19.321 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:33:19.321 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:33:19.321 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:33:19.321 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:33:19.321 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:33:19.321 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:33:19.321 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:33:19.321 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:33:19.321 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:33:19.321 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:33:19.580 /dev/nbd0 00:33:19.580 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:33:19.580 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:33:19.580 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:33:19.580 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:33:19.580 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:33:19.580 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:33:19.580 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:33:19.580 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:33:19.580 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:33:19.580 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:33:19.580 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:33:19.580 1+0 records in 00:33:19.580 1+0 records out 00:33:19.580 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000511928 s, 8.0 MB/s 00:33:19.580 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:33:19.580 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:33:19.580 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:33:19.580 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:33:19.580 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:33:19.580 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:33:19.580 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:33:19.580 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:33:19.839 /dev/nbd1 00:33:19.839 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:33:19.839 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:33:19.839 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:33:19.839 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:33:19.839 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:33:19.839 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:33:19.839 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:33:19.839 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:33:19.839 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:33:19.839 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:33:19.839 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:33:19.839 1+0 records in 00:33:19.839 1+0 records out 00:33:19.839 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000298467 s, 13.7 MB/s 00:33:19.839 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:33:19.839 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:33:19.839 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:33:19.839 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:33:19.839 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:33:19.839 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:33:19.839 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:33:19.839 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:33:19.839 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:33:19.839 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:33:19.839 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:33:19.839 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:33:19.839 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:33:19.839 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:33:19.839 13:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:33:20.113 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:33:20.113 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:33:20.113 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:33:20.113 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:33:20.113 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:33:20.113 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:33:20.113 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:33:20.113 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:33:20.113 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:33:20.113 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:20.406 [2024-11-20 13:53:27.307272] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:33:20.406 [2024-11-20 13:53:27.307346] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:20.406 [2024-11-20 13:53:27.307368] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:33:20.406 [2024-11-20 13:53:27.307380] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:20.406 [2024-11-20 13:53:27.309464] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:20.406 [2024-11-20 13:53:27.309505] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:33:20.406 [2024-11-20 13:53:27.309599] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:33:20.406 [2024-11-20 13:53:27.309647] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:33:20.406 [2024-11-20 13:53:27.309772] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:33:20.406 [2024-11-20 13:53:27.309865] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:33:20.406 spare 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:20.406 [2024-11-20 13:53:27.409966] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:33:20.406 [2024-11-20 13:53:27.410040] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:33:20.406 [2024-11-20 13:53:27.410388] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000047700 00:33:20.406 [2024-11-20 13:53:27.413437] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:33:20.406 [2024-11-20 13:53:27.413459] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:33:20.406 [2024-11-20 13:53:27.413656] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:20.406 "name": "raid_bdev1", 00:33:20.406 "uuid": "04a83dac-3571-4037-8e53-5786897ec384", 00:33:20.406 "strip_size_kb": 64, 00:33:20.406 "state": "online", 00:33:20.406 "raid_level": "raid5f", 00:33:20.406 "superblock": true, 00:33:20.406 "num_base_bdevs": 3, 00:33:20.406 "num_base_bdevs_discovered": 3, 00:33:20.406 "num_base_bdevs_operational": 3, 00:33:20.406 "base_bdevs_list": [ 00:33:20.406 { 00:33:20.406 "name": "spare", 00:33:20.406 "uuid": "06678d61-a1f4-5eb1-801a-208b14fcb645", 00:33:20.406 "is_configured": true, 00:33:20.406 "data_offset": 2048, 00:33:20.406 "data_size": 63488 00:33:20.406 }, 00:33:20.406 { 00:33:20.406 "name": "BaseBdev2", 00:33:20.406 "uuid": "45032530-0c80-5314-b7cd-d3e3d498c9e8", 00:33:20.406 "is_configured": true, 00:33:20.406 "data_offset": 2048, 00:33:20.406 "data_size": 63488 00:33:20.406 }, 00:33:20.406 { 00:33:20.406 "name": "BaseBdev3", 00:33:20.406 "uuid": "cae27a91-cd22-5c5e-8ee4-b6d0cfd7d514", 00:33:20.406 "is_configured": true, 00:33:20.406 "data_offset": 2048, 00:33:20.406 "data_size": 63488 00:33:20.406 } 00:33:20.406 ] 00:33:20.406 }' 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:20.406 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:20.973 "name": "raid_bdev1", 00:33:20.973 "uuid": "04a83dac-3571-4037-8e53-5786897ec384", 00:33:20.973 "strip_size_kb": 64, 00:33:20.973 "state": "online", 00:33:20.973 "raid_level": "raid5f", 00:33:20.973 "superblock": true, 00:33:20.973 "num_base_bdevs": 3, 00:33:20.973 "num_base_bdevs_discovered": 3, 00:33:20.973 "num_base_bdevs_operational": 3, 00:33:20.973 "base_bdevs_list": [ 00:33:20.973 { 00:33:20.973 "name": "spare", 00:33:20.973 "uuid": "06678d61-a1f4-5eb1-801a-208b14fcb645", 00:33:20.973 "is_configured": true, 00:33:20.973 "data_offset": 2048, 00:33:20.973 "data_size": 63488 00:33:20.973 }, 00:33:20.973 { 00:33:20.973 "name": "BaseBdev2", 00:33:20.973 "uuid": "45032530-0c80-5314-b7cd-d3e3d498c9e8", 00:33:20.973 "is_configured": true, 00:33:20.973 "data_offset": 2048, 00:33:20.973 "data_size": 63488 00:33:20.973 }, 00:33:20.973 { 00:33:20.973 "name": "BaseBdev3", 00:33:20.973 "uuid": "cae27a91-cd22-5c5e-8ee4-b6d0cfd7d514", 00:33:20.973 "is_configured": true, 00:33:20.973 "data_offset": 2048, 00:33:20.973 "data_size": 63488 00:33:20.973 } 00:33:20.973 ] 00:33:20.973 }' 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:20.973 [2024-11-20 13:53:27.865718] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:20.973 "name": "raid_bdev1", 00:33:20.973 "uuid": "04a83dac-3571-4037-8e53-5786897ec384", 00:33:20.973 "strip_size_kb": 64, 00:33:20.973 "state": "online", 00:33:20.973 "raid_level": "raid5f", 00:33:20.973 "superblock": true, 00:33:20.973 "num_base_bdevs": 3, 00:33:20.973 "num_base_bdevs_discovered": 2, 00:33:20.973 "num_base_bdevs_operational": 2, 00:33:20.973 "base_bdevs_list": [ 00:33:20.973 { 00:33:20.973 "name": null, 00:33:20.973 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:20.973 "is_configured": false, 00:33:20.973 "data_offset": 0, 00:33:20.973 "data_size": 63488 00:33:20.973 }, 00:33:20.973 { 00:33:20.973 "name": "BaseBdev2", 00:33:20.973 "uuid": "45032530-0c80-5314-b7cd-d3e3d498c9e8", 00:33:20.973 "is_configured": true, 00:33:20.973 "data_offset": 2048, 00:33:20.973 "data_size": 63488 00:33:20.973 }, 00:33:20.973 { 00:33:20.973 "name": "BaseBdev3", 00:33:20.973 "uuid": "cae27a91-cd22-5c5e-8ee4-b6d0cfd7d514", 00:33:20.973 "is_configured": true, 00:33:20.973 "data_offset": 2048, 00:33:20.973 "data_size": 63488 00:33:20.973 } 00:33:20.973 ] 00:33:20.973 }' 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:20.973 13:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:21.232 13:53:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:33:21.232 13:53:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:21.232 13:53:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:21.232 [2024-11-20 13:53:28.201788] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:33:21.232 [2024-11-20 13:53:28.201996] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:33:21.232 [2024-11-20 13:53:28.202018] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:33:21.232 [2024-11-20 13:53:28.202052] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:33:21.232 [2024-11-20 13:53:28.210926] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000477d0 00:33:21.232 13:53:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:21.232 13:53:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:33:21.232 [2024-11-20 13:53:28.215504] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:33:22.165 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:22.165 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:22.165 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:22.165 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:22.165 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:22.165 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:22.165 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:22.165 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:22.165 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:22.423 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:22.423 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:22.423 "name": "raid_bdev1", 00:33:22.423 "uuid": "04a83dac-3571-4037-8e53-5786897ec384", 00:33:22.423 "strip_size_kb": 64, 00:33:22.423 "state": "online", 00:33:22.423 "raid_level": "raid5f", 00:33:22.423 "superblock": true, 00:33:22.423 "num_base_bdevs": 3, 00:33:22.423 "num_base_bdevs_discovered": 3, 00:33:22.423 "num_base_bdevs_operational": 3, 00:33:22.423 "process": { 00:33:22.423 "type": "rebuild", 00:33:22.423 "target": "spare", 00:33:22.423 "progress": { 00:33:22.423 "blocks": 18432, 00:33:22.423 "percent": 14 00:33:22.423 } 00:33:22.423 }, 00:33:22.423 "base_bdevs_list": [ 00:33:22.423 { 00:33:22.423 "name": "spare", 00:33:22.423 "uuid": "06678d61-a1f4-5eb1-801a-208b14fcb645", 00:33:22.423 "is_configured": true, 00:33:22.423 "data_offset": 2048, 00:33:22.423 "data_size": 63488 00:33:22.423 }, 00:33:22.423 { 00:33:22.423 "name": "BaseBdev2", 00:33:22.423 "uuid": "45032530-0c80-5314-b7cd-d3e3d498c9e8", 00:33:22.423 "is_configured": true, 00:33:22.423 "data_offset": 2048, 00:33:22.423 "data_size": 63488 00:33:22.423 }, 00:33:22.423 { 00:33:22.423 "name": "BaseBdev3", 00:33:22.423 "uuid": "cae27a91-cd22-5c5e-8ee4-b6d0cfd7d514", 00:33:22.423 "is_configured": true, 00:33:22.423 "data_offset": 2048, 00:33:22.423 "data_size": 63488 00:33:22.423 } 00:33:22.423 ] 00:33:22.423 }' 00:33:22.423 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:22.423 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:33:22.423 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:22.423 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:33:22.423 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:33:22.423 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:22.423 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:22.423 [2024-11-20 13:53:29.321219] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:33:22.423 [2024-11-20 13:53:29.326348] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:33:22.423 [2024-11-20 13:53:29.326404] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:22.423 [2024-11-20 13:53:29.326419] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:33:22.423 [2024-11-20 13:53:29.326427] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:33:22.423 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:22.423 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:33:22.423 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:22.423 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:22.423 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:22.424 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:22.424 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:33:22.424 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:22.424 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:22.424 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:22.424 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:22.424 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:22.424 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:22.424 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:22.424 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:22.424 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:22.424 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:22.424 "name": "raid_bdev1", 00:33:22.424 "uuid": "04a83dac-3571-4037-8e53-5786897ec384", 00:33:22.424 "strip_size_kb": 64, 00:33:22.424 "state": "online", 00:33:22.424 "raid_level": "raid5f", 00:33:22.424 "superblock": true, 00:33:22.424 "num_base_bdevs": 3, 00:33:22.424 "num_base_bdevs_discovered": 2, 00:33:22.424 "num_base_bdevs_operational": 2, 00:33:22.424 "base_bdevs_list": [ 00:33:22.424 { 00:33:22.424 "name": null, 00:33:22.424 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:22.424 "is_configured": false, 00:33:22.424 "data_offset": 0, 00:33:22.424 "data_size": 63488 00:33:22.424 }, 00:33:22.424 { 00:33:22.424 "name": "BaseBdev2", 00:33:22.424 "uuid": "45032530-0c80-5314-b7cd-d3e3d498c9e8", 00:33:22.424 "is_configured": true, 00:33:22.424 "data_offset": 2048, 00:33:22.424 "data_size": 63488 00:33:22.424 }, 00:33:22.424 { 00:33:22.424 "name": "BaseBdev3", 00:33:22.424 "uuid": "cae27a91-cd22-5c5e-8ee4-b6d0cfd7d514", 00:33:22.424 "is_configured": true, 00:33:22.424 "data_offset": 2048, 00:33:22.424 "data_size": 63488 00:33:22.424 } 00:33:22.424 ] 00:33:22.424 }' 00:33:22.424 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:22.424 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:22.683 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:33:22.683 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:22.683 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:22.683 [2024-11-20 13:53:29.665866] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:33:22.683 [2024-11-20 13:53:29.665944] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:22.683 [2024-11-20 13:53:29.665966] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b780 00:33:22.683 [2024-11-20 13:53:29.666009] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:22.683 [2024-11-20 13:53:29.666461] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:22.683 [2024-11-20 13:53:29.666490] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:33:22.683 [2024-11-20 13:53:29.666581] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:33:22.683 [2024-11-20 13:53:29.666601] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:33:22.683 [2024-11-20 13:53:29.666611] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:33:22.683 [2024-11-20 13:53:29.666631] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:33:22.683 [2024-11-20 13:53:29.675271] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000478a0 00:33:22.683 spare 00:33:22.683 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:22.683 13:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:33:22.683 [2024-11-20 13:53:29.679732] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:33:23.635 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:23.635 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:23.635 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:23.635 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:23.635 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:23.635 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:23.635 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:23.635 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:23.635 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:23.894 "name": "raid_bdev1", 00:33:23.894 "uuid": "04a83dac-3571-4037-8e53-5786897ec384", 00:33:23.894 "strip_size_kb": 64, 00:33:23.894 "state": "online", 00:33:23.894 "raid_level": "raid5f", 00:33:23.894 "superblock": true, 00:33:23.894 "num_base_bdevs": 3, 00:33:23.894 "num_base_bdevs_discovered": 3, 00:33:23.894 "num_base_bdevs_operational": 3, 00:33:23.894 "process": { 00:33:23.894 "type": "rebuild", 00:33:23.894 "target": "spare", 00:33:23.894 "progress": { 00:33:23.894 "blocks": 18432, 00:33:23.894 "percent": 14 00:33:23.894 } 00:33:23.894 }, 00:33:23.894 "base_bdevs_list": [ 00:33:23.894 { 00:33:23.894 "name": "spare", 00:33:23.894 "uuid": "06678d61-a1f4-5eb1-801a-208b14fcb645", 00:33:23.894 "is_configured": true, 00:33:23.894 "data_offset": 2048, 00:33:23.894 "data_size": 63488 00:33:23.894 }, 00:33:23.894 { 00:33:23.894 "name": "BaseBdev2", 00:33:23.894 "uuid": "45032530-0c80-5314-b7cd-d3e3d498c9e8", 00:33:23.894 "is_configured": true, 00:33:23.894 "data_offset": 2048, 00:33:23.894 "data_size": 63488 00:33:23.894 }, 00:33:23.894 { 00:33:23.894 "name": "BaseBdev3", 00:33:23.894 "uuid": "cae27a91-cd22-5c5e-8ee4-b6d0cfd7d514", 00:33:23.894 "is_configured": true, 00:33:23.894 "data_offset": 2048, 00:33:23.894 "data_size": 63488 00:33:23.894 } 00:33:23.894 ] 00:33:23.894 }' 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:23.894 [2024-11-20 13:53:30.780856] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:33:23.894 [2024-11-20 13:53:30.790658] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:33:23.894 [2024-11-20 13:53:30.790728] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:23.894 [2024-11-20 13:53:30.790744] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:33:23.894 [2024-11-20 13:53:30.790751] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:23.894 "name": "raid_bdev1", 00:33:23.894 "uuid": "04a83dac-3571-4037-8e53-5786897ec384", 00:33:23.894 "strip_size_kb": 64, 00:33:23.894 "state": "online", 00:33:23.894 "raid_level": "raid5f", 00:33:23.894 "superblock": true, 00:33:23.894 "num_base_bdevs": 3, 00:33:23.894 "num_base_bdevs_discovered": 2, 00:33:23.894 "num_base_bdevs_operational": 2, 00:33:23.894 "base_bdevs_list": [ 00:33:23.894 { 00:33:23.894 "name": null, 00:33:23.894 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:23.894 "is_configured": false, 00:33:23.894 "data_offset": 0, 00:33:23.894 "data_size": 63488 00:33:23.894 }, 00:33:23.894 { 00:33:23.894 "name": "BaseBdev2", 00:33:23.894 "uuid": "45032530-0c80-5314-b7cd-d3e3d498c9e8", 00:33:23.894 "is_configured": true, 00:33:23.894 "data_offset": 2048, 00:33:23.894 "data_size": 63488 00:33:23.894 }, 00:33:23.894 { 00:33:23.894 "name": "BaseBdev3", 00:33:23.894 "uuid": "cae27a91-cd22-5c5e-8ee4-b6d0cfd7d514", 00:33:23.894 "is_configured": true, 00:33:23.894 "data_offset": 2048, 00:33:23.894 "data_size": 63488 00:33:23.894 } 00:33:23.894 ] 00:33:23.894 }' 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:23.894 13:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:24.152 13:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:33:24.152 13:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:24.152 13:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:33:24.152 13:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:33:24.152 13:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:24.152 13:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:24.152 13:53:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:24.152 13:53:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:24.152 13:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:24.152 13:53:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:24.152 13:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:24.152 "name": "raid_bdev1", 00:33:24.152 "uuid": "04a83dac-3571-4037-8e53-5786897ec384", 00:33:24.152 "strip_size_kb": 64, 00:33:24.152 "state": "online", 00:33:24.152 "raid_level": "raid5f", 00:33:24.152 "superblock": true, 00:33:24.152 "num_base_bdevs": 3, 00:33:24.152 "num_base_bdevs_discovered": 2, 00:33:24.152 "num_base_bdevs_operational": 2, 00:33:24.153 "base_bdevs_list": [ 00:33:24.153 { 00:33:24.153 "name": null, 00:33:24.153 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:24.153 "is_configured": false, 00:33:24.153 "data_offset": 0, 00:33:24.153 "data_size": 63488 00:33:24.153 }, 00:33:24.153 { 00:33:24.153 "name": "BaseBdev2", 00:33:24.153 "uuid": "45032530-0c80-5314-b7cd-d3e3d498c9e8", 00:33:24.153 "is_configured": true, 00:33:24.153 "data_offset": 2048, 00:33:24.153 "data_size": 63488 00:33:24.153 }, 00:33:24.153 { 00:33:24.153 "name": "BaseBdev3", 00:33:24.153 "uuid": "cae27a91-cd22-5c5e-8ee4-b6d0cfd7d514", 00:33:24.153 "is_configured": true, 00:33:24.153 "data_offset": 2048, 00:33:24.153 "data_size": 63488 00:33:24.153 } 00:33:24.153 ] 00:33:24.153 }' 00:33:24.153 13:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:24.153 13:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:33:24.153 13:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:24.153 13:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:33:24.153 13:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:33:24.153 13:53:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:24.153 13:53:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:24.411 13:53:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:24.412 13:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:33:24.412 13:53:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:24.412 13:53:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:24.412 [2024-11-20 13:53:31.221948] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:33:24.412 [2024-11-20 13:53:31.222014] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:24.412 [2024-11-20 13:53:31.222037] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:33:24.412 [2024-11-20 13:53:31.222045] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:24.412 [2024-11-20 13:53:31.222465] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:24.412 [2024-11-20 13:53:31.222484] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:33:24.412 [2024-11-20 13:53:31.222558] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:33:24.412 [2024-11-20 13:53:31.222573] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:33:24.412 [2024-11-20 13:53:31.222583] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:33:24.412 [2024-11-20 13:53:31.222594] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:33:24.412 BaseBdev1 00:33:24.412 13:53:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:24.412 13:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:33:25.347 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:33:25.347 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:25.347 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:25.347 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:25.347 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:25.347 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:33:25.347 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:25.347 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:25.347 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:25.347 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:25.347 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:25.347 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:25.347 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:25.347 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:25.347 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:25.347 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:25.347 "name": "raid_bdev1", 00:33:25.347 "uuid": "04a83dac-3571-4037-8e53-5786897ec384", 00:33:25.347 "strip_size_kb": 64, 00:33:25.347 "state": "online", 00:33:25.347 "raid_level": "raid5f", 00:33:25.347 "superblock": true, 00:33:25.347 "num_base_bdevs": 3, 00:33:25.347 "num_base_bdevs_discovered": 2, 00:33:25.347 "num_base_bdevs_operational": 2, 00:33:25.347 "base_bdevs_list": [ 00:33:25.347 { 00:33:25.347 "name": null, 00:33:25.347 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:25.347 "is_configured": false, 00:33:25.347 "data_offset": 0, 00:33:25.347 "data_size": 63488 00:33:25.347 }, 00:33:25.347 { 00:33:25.347 "name": "BaseBdev2", 00:33:25.347 "uuid": "45032530-0c80-5314-b7cd-d3e3d498c9e8", 00:33:25.347 "is_configured": true, 00:33:25.347 "data_offset": 2048, 00:33:25.347 "data_size": 63488 00:33:25.347 }, 00:33:25.347 { 00:33:25.347 "name": "BaseBdev3", 00:33:25.347 "uuid": "cae27a91-cd22-5c5e-8ee4-b6d0cfd7d514", 00:33:25.347 "is_configured": true, 00:33:25.347 "data_offset": 2048, 00:33:25.347 "data_size": 63488 00:33:25.347 } 00:33:25.347 ] 00:33:25.347 }' 00:33:25.347 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:25.347 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:25.604 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:33:25.604 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:25.604 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:33:25.604 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:33:25.604 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:25.604 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:25.604 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:25.604 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:25.604 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:25.604 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:25.604 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:25.604 "name": "raid_bdev1", 00:33:25.604 "uuid": "04a83dac-3571-4037-8e53-5786897ec384", 00:33:25.604 "strip_size_kb": 64, 00:33:25.604 "state": "online", 00:33:25.604 "raid_level": "raid5f", 00:33:25.604 "superblock": true, 00:33:25.604 "num_base_bdevs": 3, 00:33:25.604 "num_base_bdevs_discovered": 2, 00:33:25.604 "num_base_bdevs_operational": 2, 00:33:25.604 "base_bdevs_list": [ 00:33:25.604 { 00:33:25.604 "name": null, 00:33:25.604 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:25.604 "is_configured": false, 00:33:25.604 "data_offset": 0, 00:33:25.604 "data_size": 63488 00:33:25.604 }, 00:33:25.604 { 00:33:25.604 "name": "BaseBdev2", 00:33:25.604 "uuid": "45032530-0c80-5314-b7cd-d3e3d498c9e8", 00:33:25.604 "is_configured": true, 00:33:25.604 "data_offset": 2048, 00:33:25.604 "data_size": 63488 00:33:25.604 }, 00:33:25.604 { 00:33:25.604 "name": "BaseBdev3", 00:33:25.604 "uuid": "cae27a91-cd22-5c5e-8ee4-b6d0cfd7d514", 00:33:25.604 "is_configured": true, 00:33:25.604 "data_offset": 2048, 00:33:25.604 "data_size": 63488 00:33:25.604 } 00:33:25.604 ] 00:33:25.604 }' 00:33:25.604 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:25.604 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:33:25.604 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:25.861 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:33:25.861 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:33:25.861 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:33:25.861 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:33:25.861 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:33:25.861 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:33:25.861 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:33:25.861 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:33:25.861 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:33:25.861 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:25.861 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:25.861 [2024-11-20 13:53:32.694310] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:33:25.861 [2024-11-20 13:53:32.694481] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:33:25.861 [2024-11-20 13:53:32.694494] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:33:25.861 request: 00:33:25.861 { 00:33:25.861 "base_bdev": "BaseBdev1", 00:33:25.861 "raid_bdev": "raid_bdev1", 00:33:25.861 "method": "bdev_raid_add_base_bdev", 00:33:25.861 "req_id": 1 00:33:25.861 } 00:33:25.861 Got JSON-RPC error response 00:33:25.861 response: 00:33:25.861 { 00:33:25.861 "code": -22, 00:33:25.861 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:33:25.861 } 00:33:25.861 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:33:25.861 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:33:25.861 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:33:25.861 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:33:25.861 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:33:25.861 13:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:33:26.847 13:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:33:26.847 13:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:26.847 13:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:26.847 13:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:26.847 13:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:26.847 13:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:33:26.847 13:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:26.847 13:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:26.847 13:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:26.847 13:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:26.847 13:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:26.847 13:53:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:26.847 13:53:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:26.847 13:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:26.847 13:53:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:26.847 13:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:26.847 "name": "raid_bdev1", 00:33:26.847 "uuid": "04a83dac-3571-4037-8e53-5786897ec384", 00:33:26.847 "strip_size_kb": 64, 00:33:26.847 "state": "online", 00:33:26.847 "raid_level": "raid5f", 00:33:26.847 "superblock": true, 00:33:26.847 "num_base_bdevs": 3, 00:33:26.847 "num_base_bdevs_discovered": 2, 00:33:26.847 "num_base_bdevs_operational": 2, 00:33:26.847 "base_bdevs_list": [ 00:33:26.847 { 00:33:26.847 "name": null, 00:33:26.847 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:26.847 "is_configured": false, 00:33:26.847 "data_offset": 0, 00:33:26.847 "data_size": 63488 00:33:26.847 }, 00:33:26.847 { 00:33:26.847 "name": "BaseBdev2", 00:33:26.847 "uuid": "45032530-0c80-5314-b7cd-d3e3d498c9e8", 00:33:26.847 "is_configured": true, 00:33:26.847 "data_offset": 2048, 00:33:26.847 "data_size": 63488 00:33:26.847 }, 00:33:26.847 { 00:33:26.847 "name": "BaseBdev3", 00:33:26.847 "uuid": "cae27a91-cd22-5c5e-8ee4-b6d0cfd7d514", 00:33:26.847 "is_configured": true, 00:33:26.847 "data_offset": 2048, 00:33:26.847 "data_size": 63488 00:33:26.847 } 00:33:26.847 ] 00:33:26.847 }' 00:33:26.847 13:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:26.847 13:53:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:27.105 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:33:27.105 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:27.105 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:33:27.105 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:33:27.105 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:27.105 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:27.105 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:27.105 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:27.105 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:27.105 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:27.105 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:27.105 "name": "raid_bdev1", 00:33:27.105 "uuid": "04a83dac-3571-4037-8e53-5786897ec384", 00:33:27.105 "strip_size_kb": 64, 00:33:27.105 "state": "online", 00:33:27.105 "raid_level": "raid5f", 00:33:27.105 "superblock": true, 00:33:27.105 "num_base_bdevs": 3, 00:33:27.105 "num_base_bdevs_discovered": 2, 00:33:27.105 "num_base_bdevs_operational": 2, 00:33:27.105 "base_bdevs_list": [ 00:33:27.105 { 00:33:27.105 "name": null, 00:33:27.105 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:27.105 "is_configured": false, 00:33:27.106 "data_offset": 0, 00:33:27.106 "data_size": 63488 00:33:27.106 }, 00:33:27.106 { 00:33:27.106 "name": "BaseBdev2", 00:33:27.106 "uuid": "45032530-0c80-5314-b7cd-d3e3d498c9e8", 00:33:27.106 "is_configured": true, 00:33:27.106 "data_offset": 2048, 00:33:27.106 "data_size": 63488 00:33:27.106 }, 00:33:27.106 { 00:33:27.106 "name": "BaseBdev3", 00:33:27.106 "uuid": "cae27a91-cd22-5c5e-8ee4-b6d0cfd7d514", 00:33:27.106 "is_configured": true, 00:33:27.106 "data_offset": 2048, 00:33:27.106 "data_size": 63488 00:33:27.106 } 00:33:27.106 ] 00:33:27.106 }' 00:33:27.106 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:27.106 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:33:27.106 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:27.106 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:33:27.106 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 79660 00:33:27.106 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 79660 ']' 00:33:27.106 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 79660 00:33:27.106 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:33:27.106 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:33:27.106 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 79660 00:33:27.106 killing process with pid 79660 00:33:27.106 Received shutdown signal, test time was about 60.000000 seconds 00:33:27.106 00:33:27.106 Latency(us) 00:33:27.106 [2024-11-20T13:53:34.165Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:33:27.106 [2024-11-20T13:53:34.165Z] =================================================================================================================== 00:33:27.106 [2024-11-20T13:53:34.165Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:33:27.106 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:33:27.106 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:33:27.106 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 79660' 00:33:27.106 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 79660 00:33:27.106 [2024-11-20 13:53:34.151695] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:33:27.106 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 79660 00:33:27.106 [2024-11-20 13:53:34.151815] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:33:27.106 [2024-11-20 13:53:34.151878] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:33:27.106 [2024-11-20 13:53:34.151888] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:33:27.363 [2024-11-20 13:53:34.355311] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:33:27.930 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:33:27.930 00:33:27.930 real 0m20.143s 00:33:27.930 user 0m25.065s 00:33:27.930 sys 0m2.106s 00:33:27.930 ************************************ 00:33:27.930 END TEST raid5f_rebuild_test_sb 00:33:27.930 ************************************ 00:33:27.930 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:33:27.930 13:53:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:27.931 13:53:34 bdev_raid -- bdev/bdev_raid.sh@985 -- # for n in {3..4} 00:33:27.931 13:53:34 bdev_raid -- bdev/bdev_raid.sh@986 -- # run_test raid5f_state_function_test raid_state_function_test raid5f 4 false 00:33:27.931 13:53:34 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:33:27.931 13:53:34 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:33:27.931 13:53:34 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:33:28.189 ************************************ 00:33:28.189 START TEST raid5f_state_function_test 00:33:28.189 ************************************ 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 4 false 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:33:28.189 Process raid pid: 80385 00:33:28.189 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=80385 00:33:28.189 13:53:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 80385' 00:33:28.189 13:53:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:33:28.189 13:53:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 80385 00:33:28.189 13:53:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 80385 ']' 00:33:28.189 13:53:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:33:28.189 13:53:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:33:28.189 13:53:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:33:28.189 13:53:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:33:28.189 13:53:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:28.189 [2024-11-20 13:53:35.082872] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:33:28.189 [2024-11-20 13:53:35.083065] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:33:28.447 [2024-11-20 13:53:35.258596] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:33:28.447 [2024-11-20 13:53:35.380138] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:33:28.705 [2024-11-20 13:53:35.530640] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:33:28.705 [2024-11-20 13:53:35.530691] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:33:28.964 13:53:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:33:28.964 13:53:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:33:28.964 13:53:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:33:28.964 13:53:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:28.964 13:53:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:28.964 [2024-11-20 13:53:35.867973] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:33:28.964 [2024-11-20 13:53:35.868067] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:33:28.964 [2024-11-20 13:53:35.868085] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:33:28.964 [2024-11-20 13:53:35.868100] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:33:28.964 [2024-11-20 13:53:35.868112] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:33:28.964 [2024-11-20 13:53:35.868127] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:33:28.964 [2024-11-20 13:53:35.868138] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:33:28.964 [2024-11-20 13:53:35.868152] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:33:28.964 13:53:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:28.964 13:53:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:33:28.964 13:53:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:28.964 13:53:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:28.964 13:53:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:28.964 13:53:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:28.964 13:53:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:28.964 13:53:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:28.964 13:53:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:28.964 13:53:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:28.964 13:53:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:28.964 13:53:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:28.964 13:53:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:28.964 13:53:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:28.964 13:53:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:28.964 13:53:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:28.964 13:53:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:28.964 "name": "Existed_Raid", 00:33:28.964 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:28.964 "strip_size_kb": 64, 00:33:28.964 "state": "configuring", 00:33:28.964 "raid_level": "raid5f", 00:33:28.964 "superblock": false, 00:33:28.964 "num_base_bdevs": 4, 00:33:28.964 "num_base_bdevs_discovered": 0, 00:33:28.964 "num_base_bdevs_operational": 4, 00:33:28.964 "base_bdevs_list": [ 00:33:28.964 { 00:33:28.964 "name": "BaseBdev1", 00:33:28.964 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:28.964 "is_configured": false, 00:33:28.964 "data_offset": 0, 00:33:28.964 "data_size": 0 00:33:28.964 }, 00:33:28.964 { 00:33:28.964 "name": "BaseBdev2", 00:33:28.964 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:28.964 "is_configured": false, 00:33:28.964 "data_offset": 0, 00:33:28.964 "data_size": 0 00:33:28.964 }, 00:33:28.964 { 00:33:28.964 "name": "BaseBdev3", 00:33:28.964 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:28.964 "is_configured": false, 00:33:28.964 "data_offset": 0, 00:33:28.964 "data_size": 0 00:33:28.964 }, 00:33:28.964 { 00:33:28.964 "name": "BaseBdev4", 00:33:28.964 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:28.964 "is_configured": false, 00:33:28.964 "data_offset": 0, 00:33:28.964 "data_size": 0 00:33:28.964 } 00:33:28.964 ] 00:33:28.964 }' 00:33:28.964 13:53:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:28.964 13:53:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:29.222 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:33:29.222 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:29.222 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:29.222 [2024-11-20 13:53:36.199943] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:33:29.222 [2024-11-20 13:53:36.200008] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:33:29.222 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:29.222 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:29.223 [2024-11-20 13:53:36.207947] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:33:29.223 [2024-11-20 13:53:36.208005] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:33:29.223 [2024-11-20 13:53:36.208016] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:33:29.223 [2024-11-20 13:53:36.208026] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:33:29.223 [2024-11-20 13:53:36.208033] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:33:29.223 [2024-11-20 13:53:36.208042] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:33:29.223 [2024-11-20 13:53:36.208048] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:33:29.223 [2024-11-20 13:53:36.208057] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:29.223 [2024-11-20 13:53:36.242861] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:33:29.223 BaseBdev1 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:29.223 [ 00:33:29.223 { 00:33:29.223 "name": "BaseBdev1", 00:33:29.223 "aliases": [ 00:33:29.223 "0f02072d-992a-40e3-a003-b9984798fc83" 00:33:29.223 ], 00:33:29.223 "product_name": "Malloc disk", 00:33:29.223 "block_size": 512, 00:33:29.223 "num_blocks": 65536, 00:33:29.223 "uuid": "0f02072d-992a-40e3-a003-b9984798fc83", 00:33:29.223 "assigned_rate_limits": { 00:33:29.223 "rw_ios_per_sec": 0, 00:33:29.223 "rw_mbytes_per_sec": 0, 00:33:29.223 "r_mbytes_per_sec": 0, 00:33:29.223 "w_mbytes_per_sec": 0 00:33:29.223 }, 00:33:29.223 "claimed": true, 00:33:29.223 "claim_type": "exclusive_write", 00:33:29.223 "zoned": false, 00:33:29.223 "supported_io_types": { 00:33:29.223 "read": true, 00:33:29.223 "write": true, 00:33:29.223 "unmap": true, 00:33:29.223 "flush": true, 00:33:29.223 "reset": true, 00:33:29.223 "nvme_admin": false, 00:33:29.223 "nvme_io": false, 00:33:29.223 "nvme_io_md": false, 00:33:29.223 "write_zeroes": true, 00:33:29.223 "zcopy": true, 00:33:29.223 "get_zone_info": false, 00:33:29.223 "zone_management": false, 00:33:29.223 "zone_append": false, 00:33:29.223 "compare": false, 00:33:29.223 "compare_and_write": false, 00:33:29.223 "abort": true, 00:33:29.223 "seek_hole": false, 00:33:29.223 "seek_data": false, 00:33:29.223 "copy": true, 00:33:29.223 "nvme_iov_md": false 00:33:29.223 }, 00:33:29.223 "memory_domains": [ 00:33:29.223 { 00:33:29.223 "dma_device_id": "system", 00:33:29.223 "dma_device_type": 1 00:33:29.223 }, 00:33:29.223 { 00:33:29.223 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:29.223 "dma_device_type": 2 00:33:29.223 } 00:33:29.223 ], 00:33:29.223 "driver_specific": {} 00:33:29.223 } 00:33:29.223 ] 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:29.223 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:29.482 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:29.482 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:29.482 "name": "Existed_Raid", 00:33:29.482 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:29.482 "strip_size_kb": 64, 00:33:29.482 "state": "configuring", 00:33:29.482 "raid_level": "raid5f", 00:33:29.482 "superblock": false, 00:33:29.482 "num_base_bdevs": 4, 00:33:29.482 "num_base_bdevs_discovered": 1, 00:33:29.482 "num_base_bdevs_operational": 4, 00:33:29.482 "base_bdevs_list": [ 00:33:29.482 { 00:33:29.482 "name": "BaseBdev1", 00:33:29.482 "uuid": "0f02072d-992a-40e3-a003-b9984798fc83", 00:33:29.482 "is_configured": true, 00:33:29.482 "data_offset": 0, 00:33:29.482 "data_size": 65536 00:33:29.482 }, 00:33:29.482 { 00:33:29.482 "name": "BaseBdev2", 00:33:29.482 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:29.482 "is_configured": false, 00:33:29.482 "data_offset": 0, 00:33:29.482 "data_size": 0 00:33:29.482 }, 00:33:29.482 { 00:33:29.482 "name": "BaseBdev3", 00:33:29.482 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:29.482 "is_configured": false, 00:33:29.482 "data_offset": 0, 00:33:29.482 "data_size": 0 00:33:29.482 }, 00:33:29.482 { 00:33:29.482 "name": "BaseBdev4", 00:33:29.482 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:29.482 "is_configured": false, 00:33:29.482 "data_offset": 0, 00:33:29.482 "data_size": 0 00:33:29.482 } 00:33:29.482 ] 00:33:29.482 }' 00:33:29.482 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:29.482 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:29.740 [2024-11-20 13:53:36.611034] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:33:29.740 [2024-11-20 13:53:36.611094] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:29.740 [2024-11-20 13:53:36.619074] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:33:29.740 [2024-11-20 13:53:36.621076] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:33:29.740 [2024-11-20 13:53:36.621119] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:33:29.740 [2024-11-20 13:53:36.621129] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:33:29.740 [2024-11-20 13:53:36.621142] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:33:29.740 [2024-11-20 13:53:36.621149] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:33:29.740 [2024-11-20 13:53:36.621158] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:29.740 "name": "Existed_Raid", 00:33:29.740 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:29.740 "strip_size_kb": 64, 00:33:29.740 "state": "configuring", 00:33:29.740 "raid_level": "raid5f", 00:33:29.740 "superblock": false, 00:33:29.740 "num_base_bdevs": 4, 00:33:29.740 "num_base_bdevs_discovered": 1, 00:33:29.740 "num_base_bdevs_operational": 4, 00:33:29.740 "base_bdevs_list": [ 00:33:29.740 { 00:33:29.740 "name": "BaseBdev1", 00:33:29.740 "uuid": "0f02072d-992a-40e3-a003-b9984798fc83", 00:33:29.740 "is_configured": true, 00:33:29.740 "data_offset": 0, 00:33:29.740 "data_size": 65536 00:33:29.740 }, 00:33:29.740 { 00:33:29.740 "name": "BaseBdev2", 00:33:29.740 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:29.740 "is_configured": false, 00:33:29.740 "data_offset": 0, 00:33:29.740 "data_size": 0 00:33:29.740 }, 00:33:29.740 { 00:33:29.740 "name": "BaseBdev3", 00:33:29.740 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:29.740 "is_configured": false, 00:33:29.740 "data_offset": 0, 00:33:29.740 "data_size": 0 00:33:29.740 }, 00:33:29.740 { 00:33:29.740 "name": "BaseBdev4", 00:33:29.740 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:29.740 "is_configured": false, 00:33:29.740 "data_offset": 0, 00:33:29.740 "data_size": 0 00:33:29.740 } 00:33:29.740 ] 00:33:29.740 }' 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:29.740 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:30.076 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:33:30.076 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:30.076 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:30.076 [2024-11-20 13:53:36.983812] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:33:30.076 BaseBdev2 00:33:30.076 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:30.076 13:53:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:33:30.076 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:33:30.076 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:33:30.076 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:33:30.076 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:33:30.076 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:33:30.076 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:33:30.076 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:30.076 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:30.076 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:30.077 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:33:30.077 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:30.077 13:53:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:30.077 [ 00:33:30.077 { 00:33:30.077 "name": "BaseBdev2", 00:33:30.077 "aliases": [ 00:33:30.077 "32b29861-955e-4271-90c4-3ea512864a27" 00:33:30.077 ], 00:33:30.077 "product_name": "Malloc disk", 00:33:30.077 "block_size": 512, 00:33:30.077 "num_blocks": 65536, 00:33:30.077 "uuid": "32b29861-955e-4271-90c4-3ea512864a27", 00:33:30.077 "assigned_rate_limits": { 00:33:30.077 "rw_ios_per_sec": 0, 00:33:30.077 "rw_mbytes_per_sec": 0, 00:33:30.077 "r_mbytes_per_sec": 0, 00:33:30.077 "w_mbytes_per_sec": 0 00:33:30.077 }, 00:33:30.077 "claimed": true, 00:33:30.077 "claim_type": "exclusive_write", 00:33:30.077 "zoned": false, 00:33:30.077 "supported_io_types": { 00:33:30.077 "read": true, 00:33:30.077 "write": true, 00:33:30.077 "unmap": true, 00:33:30.077 "flush": true, 00:33:30.077 "reset": true, 00:33:30.077 "nvme_admin": false, 00:33:30.077 "nvme_io": false, 00:33:30.077 "nvme_io_md": false, 00:33:30.077 "write_zeroes": true, 00:33:30.077 "zcopy": true, 00:33:30.077 "get_zone_info": false, 00:33:30.077 "zone_management": false, 00:33:30.077 "zone_append": false, 00:33:30.077 "compare": false, 00:33:30.077 "compare_and_write": false, 00:33:30.077 "abort": true, 00:33:30.077 "seek_hole": false, 00:33:30.077 "seek_data": false, 00:33:30.077 "copy": true, 00:33:30.077 "nvme_iov_md": false 00:33:30.077 }, 00:33:30.077 "memory_domains": [ 00:33:30.077 { 00:33:30.077 "dma_device_id": "system", 00:33:30.077 "dma_device_type": 1 00:33:30.077 }, 00:33:30.077 { 00:33:30.077 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:30.077 "dma_device_type": 2 00:33:30.077 } 00:33:30.077 ], 00:33:30.077 "driver_specific": {} 00:33:30.077 } 00:33:30.077 ] 00:33:30.077 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:30.077 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:33:30.077 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:33:30.077 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:33:30.077 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:33:30.077 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:30.077 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:30.077 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:30.077 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:30.077 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:30.077 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:30.077 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:30.077 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:30.077 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:30.077 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:30.077 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:30.077 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:30.077 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:30.077 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:30.077 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:30.077 "name": "Existed_Raid", 00:33:30.077 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:30.077 "strip_size_kb": 64, 00:33:30.077 "state": "configuring", 00:33:30.077 "raid_level": "raid5f", 00:33:30.077 "superblock": false, 00:33:30.077 "num_base_bdevs": 4, 00:33:30.077 "num_base_bdevs_discovered": 2, 00:33:30.077 "num_base_bdevs_operational": 4, 00:33:30.077 "base_bdevs_list": [ 00:33:30.077 { 00:33:30.077 "name": "BaseBdev1", 00:33:30.077 "uuid": "0f02072d-992a-40e3-a003-b9984798fc83", 00:33:30.077 "is_configured": true, 00:33:30.077 "data_offset": 0, 00:33:30.077 "data_size": 65536 00:33:30.077 }, 00:33:30.077 { 00:33:30.077 "name": "BaseBdev2", 00:33:30.077 "uuid": "32b29861-955e-4271-90c4-3ea512864a27", 00:33:30.077 "is_configured": true, 00:33:30.077 "data_offset": 0, 00:33:30.077 "data_size": 65536 00:33:30.077 }, 00:33:30.077 { 00:33:30.077 "name": "BaseBdev3", 00:33:30.077 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:30.077 "is_configured": false, 00:33:30.077 "data_offset": 0, 00:33:30.077 "data_size": 0 00:33:30.077 }, 00:33:30.077 { 00:33:30.077 "name": "BaseBdev4", 00:33:30.077 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:30.077 "is_configured": false, 00:33:30.077 "data_offset": 0, 00:33:30.077 "data_size": 0 00:33:30.077 } 00:33:30.077 ] 00:33:30.077 }' 00:33:30.077 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:30.077 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:30.336 [2024-11-20 13:53:37.346811] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:33:30.336 BaseBdev3 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:30.336 [ 00:33:30.336 { 00:33:30.336 "name": "BaseBdev3", 00:33:30.336 "aliases": [ 00:33:30.336 "dd2ef8c5-bf2b-4733-b57e-2dfb68c5edf1" 00:33:30.336 ], 00:33:30.336 "product_name": "Malloc disk", 00:33:30.336 "block_size": 512, 00:33:30.336 "num_blocks": 65536, 00:33:30.336 "uuid": "dd2ef8c5-bf2b-4733-b57e-2dfb68c5edf1", 00:33:30.336 "assigned_rate_limits": { 00:33:30.336 "rw_ios_per_sec": 0, 00:33:30.336 "rw_mbytes_per_sec": 0, 00:33:30.336 "r_mbytes_per_sec": 0, 00:33:30.336 "w_mbytes_per_sec": 0 00:33:30.336 }, 00:33:30.336 "claimed": true, 00:33:30.336 "claim_type": "exclusive_write", 00:33:30.336 "zoned": false, 00:33:30.336 "supported_io_types": { 00:33:30.336 "read": true, 00:33:30.336 "write": true, 00:33:30.336 "unmap": true, 00:33:30.336 "flush": true, 00:33:30.336 "reset": true, 00:33:30.336 "nvme_admin": false, 00:33:30.336 "nvme_io": false, 00:33:30.336 "nvme_io_md": false, 00:33:30.336 "write_zeroes": true, 00:33:30.336 "zcopy": true, 00:33:30.336 "get_zone_info": false, 00:33:30.336 "zone_management": false, 00:33:30.336 "zone_append": false, 00:33:30.336 "compare": false, 00:33:30.336 "compare_and_write": false, 00:33:30.336 "abort": true, 00:33:30.336 "seek_hole": false, 00:33:30.336 "seek_data": false, 00:33:30.336 "copy": true, 00:33:30.336 "nvme_iov_md": false 00:33:30.336 }, 00:33:30.336 "memory_domains": [ 00:33:30.336 { 00:33:30.336 "dma_device_id": "system", 00:33:30.336 "dma_device_type": 1 00:33:30.336 }, 00:33:30.336 { 00:33:30.336 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:30.336 "dma_device_type": 2 00:33:30.336 } 00:33:30.336 ], 00:33:30.336 "driver_specific": {} 00:33:30.336 } 00:33:30.336 ] 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:30.336 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:30.597 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:30.598 "name": "Existed_Raid", 00:33:30.598 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:30.598 "strip_size_kb": 64, 00:33:30.598 "state": "configuring", 00:33:30.598 "raid_level": "raid5f", 00:33:30.598 "superblock": false, 00:33:30.598 "num_base_bdevs": 4, 00:33:30.598 "num_base_bdevs_discovered": 3, 00:33:30.598 "num_base_bdevs_operational": 4, 00:33:30.598 "base_bdevs_list": [ 00:33:30.598 { 00:33:30.598 "name": "BaseBdev1", 00:33:30.598 "uuid": "0f02072d-992a-40e3-a003-b9984798fc83", 00:33:30.598 "is_configured": true, 00:33:30.598 "data_offset": 0, 00:33:30.598 "data_size": 65536 00:33:30.598 }, 00:33:30.598 { 00:33:30.598 "name": "BaseBdev2", 00:33:30.598 "uuid": "32b29861-955e-4271-90c4-3ea512864a27", 00:33:30.598 "is_configured": true, 00:33:30.598 "data_offset": 0, 00:33:30.598 "data_size": 65536 00:33:30.598 }, 00:33:30.598 { 00:33:30.598 "name": "BaseBdev3", 00:33:30.598 "uuid": "dd2ef8c5-bf2b-4733-b57e-2dfb68c5edf1", 00:33:30.598 "is_configured": true, 00:33:30.598 "data_offset": 0, 00:33:30.598 "data_size": 65536 00:33:30.598 }, 00:33:30.598 { 00:33:30.598 "name": "BaseBdev4", 00:33:30.598 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:30.598 "is_configured": false, 00:33:30.598 "data_offset": 0, 00:33:30.598 "data_size": 0 00:33:30.598 } 00:33:30.598 ] 00:33:30.598 }' 00:33:30.598 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:30.598 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:30.857 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:33:30.857 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:30.858 [2024-11-20 13:53:37.731610] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:33:30.858 [2024-11-20 13:53:37.731815] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:33:30.858 [2024-11-20 13:53:37.731848] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:33:30.858 [2024-11-20 13:53:37.732227] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:33:30.858 [2024-11-20 13:53:37.737322] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:33:30.858 [2024-11-20 13:53:37.737421] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:33:30.858 [2024-11-20 13:53:37.737758] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:30.858 BaseBdev4 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:30.858 [ 00:33:30.858 { 00:33:30.858 "name": "BaseBdev4", 00:33:30.858 "aliases": [ 00:33:30.858 "05b631d5-46e5-4301-be11-009c539de9cf" 00:33:30.858 ], 00:33:30.858 "product_name": "Malloc disk", 00:33:30.858 "block_size": 512, 00:33:30.858 "num_blocks": 65536, 00:33:30.858 "uuid": "05b631d5-46e5-4301-be11-009c539de9cf", 00:33:30.858 "assigned_rate_limits": { 00:33:30.858 "rw_ios_per_sec": 0, 00:33:30.858 "rw_mbytes_per_sec": 0, 00:33:30.858 "r_mbytes_per_sec": 0, 00:33:30.858 "w_mbytes_per_sec": 0 00:33:30.858 }, 00:33:30.858 "claimed": true, 00:33:30.858 "claim_type": "exclusive_write", 00:33:30.858 "zoned": false, 00:33:30.858 "supported_io_types": { 00:33:30.858 "read": true, 00:33:30.858 "write": true, 00:33:30.858 "unmap": true, 00:33:30.858 "flush": true, 00:33:30.858 "reset": true, 00:33:30.858 "nvme_admin": false, 00:33:30.858 "nvme_io": false, 00:33:30.858 "nvme_io_md": false, 00:33:30.858 "write_zeroes": true, 00:33:30.858 "zcopy": true, 00:33:30.858 "get_zone_info": false, 00:33:30.858 "zone_management": false, 00:33:30.858 "zone_append": false, 00:33:30.858 "compare": false, 00:33:30.858 "compare_and_write": false, 00:33:30.858 "abort": true, 00:33:30.858 "seek_hole": false, 00:33:30.858 "seek_data": false, 00:33:30.858 "copy": true, 00:33:30.858 "nvme_iov_md": false 00:33:30.858 }, 00:33:30.858 "memory_domains": [ 00:33:30.858 { 00:33:30.858 "dma_device_id": "system", 00:33:30.858 "dma_device_type": 1 00:33:30.858 }, 00:33:30.858 { 00:33:30.858 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:30.858 "dma_device_type": 2 00:33:30.858 } 00:33:30.858 ], 00:33:30.858 "driver_specific": {} 00:33:30.858 } 00:33:30.858 ] 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:30.858 "name": "Existed_Raid", 00:33:30.858 "uuid": "79ea338a-fb90-4d91-a5b7-c8acb272a013", 00:33:30.858 "strip_size_kb": 64, 00:33:30.858 "state": "online", 00:33:30.858 "raid_level": "raid5f", 00:33:30.858 "superblock": false, 00:33:30.858 "num_base_bdevs": 4, 00:33:30.858 "num_base_bdevs_discovered": 4, 00:33:30.858 "num_base_bdevs_operational": 4, 00:33:30.858 "base_bdevs_list": [ 00:33:30.858 { 00:33:30.858 "name": "BaseBdev1", 00:33:30.858 "uuid": "0f02072d-992a-40e3-a003-b9984798fc83", 00:33:30.858 "is_configured": true, 00:33:30.858 "data_offset": 0, 00:33:30.858 "data_size": 65536 00:33:30.858 }, 00:33:30.858 { 00:33:30.858 "name": "BaseBdev2", 00:33:30.858 "uuid": "32b29861-955e-4271-90c4-3ea512864a27", 00:33:30.858 "is_configured": true, 00:33:30.858 "data_offset": 0, 00:33:30.858 "data_size": 65536 00:33:30.858 }, 00:33:30.858 { 00:33:30.858 "name": "BaseBdev3", 00:33:30.858 "uuid": "dd2ef8c5-bf2b-4733-b57e-2dfb68c5edf1", 00:33:30.858 "is_configured": true, 00:33:30.858 "data_offset": 0, 00:33:30.858 "data_size": 65536 00:33:30.858 }, 00:33:30.858 { 00:33:30.858 "name": "BaseBdev4", 00:33:30.858 "uuid": "05b631d5-46e5-4301-be11-009c539de9cf", 00:33:30.858 "is_configured": true, 00:33:30.858 "data_offset": 0, 00:33:30.858 "data_size": 65536 00:33:30.858 } 00:33:30.858 ] 00:33:30.858 }' 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:30.858 13:53:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:31.117 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:33:31.117 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:33:31.117 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:33:31.117 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:33:31.117 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:33:31.117 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:33:31.117 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:33:31.117 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:31.117 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:31.117 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:33:31.117 [2024-11-20 13:53:38.091676] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:33:31.117 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:31.117 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:33:31.117 "name": "Existed_Raid", 00:33:31.117 "aliases": [ 00:33:31.117 "79ea338a-fb90-4d91-a5b7-c8acb272a013" 00:33:31.117 ], 00:33:31.117 "product_name": "Raid Volume", 00:33:31.117 "block_size": 512, 00:33:31.117 "num_blocks": 196608, 00:33:31.117 "uuid": "79ea338a-fb90-4d91-a5b7-c8acb272a013", 00:33:31.117 "assigned_rate_limits": { 00:33:31.117 "rw_ios_per_sec": 0, 00:33:31.117 "rw_mbytes_per_sec": 0, 00:33:31.117 "r_mbytes_per_sec": 0, 00:33:31.117 "w_mbytes_per_sec": 0 00:33:31.117 }, 00:33:31.117 "claimed": false, 00:33:31.117 "zoned": false, 00:33:31.117 "supported_io_types": { 00:33:31.117 "read": true, 00:33:31.117 "write": true, 00:33:31.117 "unmap": false, 00:33:31.117 "flush": false, 00:33:31.117 "reset": true, 00:33:31.117 "nvme_admin": false, 00:33:31.117 "nvme_io": false, 00:33:31.117 "nvme_io_md": false, 00:33:31.117 "write_zeroes": true, 00:33:31.117 "zcopy": false, 00:33:31.117 "get_zone_info": false, 00:33:31.117 "zone_management": false, 00:33:31.117 "zone_append": false, 00:33:31.117 "compare": false, 00:33:31.117 "compare_and_write": false, 00:33:31.117 "abort": false, 00:33:31.117 "seek_hole": false, 00:33:31.117 "seek_data": false, 00:33:31.117 "copy": false, 00:33:31.117 "nvme_iov_md": false 00:33:31.117 }, 00:33:31.117 "driver_specific": { 00:33:31.117 "raid": { 00:33:31.117 "uuid": "79ea338a-fb90-4d91-a5b7-c8acb272a013", 00:33:31.117 "strip_size_kb": 64, 00:33:31.117 "state": "online", 00:33:31.117 "raid_level": "raid5f", 00:33:31.117 "superblock": false, 00:33:31.117 "num_base_bdevs": 4, 00:33:31.117 "num_base_bdevs_discovered": 4, 00:33:31.117 "num_base_bdevs_operational": 4, 00:33:31.117 "base_bdevs_list": [ 00:33:31.117 { 00:33:31.117 "name": "BaseBdev1", 00:33:31.117 "uuid": "0f02072d-992a-40e3-a003-b9984798fc83", 00:33:31.117 "is_configured": true, 00:33:31.117 "data_offset": 0, 00:33:31.117 "data_size": 65536 00:33:31.117 }, 00:33:31.117 { 00:33:31.117 "name": "BaseBdev2", 00:33:31.117 "uuid": "32b29861-955e-4271-90c4-3ea512864a27", 00:33:31.117 "is_configured": true, 00:33:31.117 "data_offset": 0, 00:33:31.117 "data_size": 65536 00:33:31.117 }, 00:33:31.117 { 00:33:31.117 "name": "BaseBdev3", 00:33:31.117 "uuid": "dd2ef8c5-bf2b-4733-b57e-2dfb68c5edf1", 00:33:31.117 "is_configured": true, 00:33:31.117 "data_offset": 0, 00:33:31.117 "data_size": 65536 00:33:31.117 }, 00:33:31.117 { 00:33:31.118 "name": "BaseBdev4", 00:33:31.118 "uuid": "05b631d5-46e5-4301-be11-009c539de9cf", 00:33:31.118 "is_configured": true, 00:33:31.118 "data_offset": 0, 00:33:31.118 "data_size": 65536 00:33:31.118 } 00:33:31.118 ] 00:33:31.118 } 00:33:31.118 } 00:33:31.118 }' 00:33:31.118 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:33:31.118 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:33:31.118 BaseBdev2 00:33:31.118 BaseBdev3 00:33:31.118 BaseBdev4' 00:33:31.118 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:31.378 [2024-11-20 13:53:38.323512] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:31.378 "name": "Existed_Raid", 00:33:31.378 "uuid": "79ea338a-fb90-4d91-a5b7-c8acb272a013", 00:33:31.378 "strip_size_kb": 64, 00:33:31.378 "state": "online", 00:33:31.378 "raid_level": "raid5f", 00:33:31.378 "superblock": false, 00:33:31.378 "num_base_bdevs": 4, 00:33:31.378 "num_base_bdevs_discovered": 3, 00:33:31.378 "num_base_bdevs_operational": 3, 00:33:31.378 "base_bdevs_list": [ 00:33:31.378 { 00:33:31.378 "name": null, 00:33:31.378 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:31.378 "is_configured": false, 00:33:31.378 "data_offset": 0, 00:33:31.378 "data_size": 65536 00:33:31.378 }, 00:33:31.378 { 00:33:31.378 "name": "BaseBdev2", 00:33:31.378 "uuid": "32b29861-955e-4271-90c4-3ea512864a27", 00:33:31.378 "is_configured": true, 00:33:31.378 "data_offset": 0, 00:33:31.378 "data_size": 65536 00:33:31.378 }, 00:33:31.378 { 00:33:31.378 "name": "BaseBdev3", 00:33:31.378 "uuid": "dd2ef8c5-bf2b-4733-b57e-2dfb68c5edf1", 00:33:31.378 "is_configured": true, 00:33:31.378 "data_offset": 0, 00:33:31.378 "data_size": 65536 00:33:31.378 }, 00:33:31.378 { 00:33:31.378 "name": "BaseBdev4", 00:33:31.378 "uuid": "05b631d5-46e5-4301-be11-009c539de9cf", 00:33:31.378 "is_configured": true, 00:33:31.378 "data_offset": 0, 00:33:31.378 "data_size": 65536 00:33:31.378 } 00:33:31.378 ] 00:33:31.378 }' 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:31.378 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:31.948 [2024-11-20 13:53:38.777627] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:33:31.948 [2024-11-20 13:53:38.777746] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:33:31.948 [2024-11-20 13:53:38.840522] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:31.948 [2024-11-20 13:53:38.880589] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:31.948 13:53:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:31.948 [2024-11-20 13:53:38.983882] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:33:31.948 [2024-11-20 13:53:38.984073] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:33:32.208 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:32.208 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:33:32.208 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:33:32.208 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:32.208 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:33:32.208 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:32.208 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:32.208 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:32.208 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:33:32.208 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:33:32.208 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:33:32.208 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:33:32.208 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:33:32.208 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:33:32.208 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:32.208 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:32.208 BaseBdev2 00:33:32.208 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:32.208 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:33:32.208 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:33:32.208 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:33:32.208 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:33:32.208 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:33:32.208 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:33:32.208 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:33:32.208 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:32.208 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:32.209 [ 00:33:32.209 { 00:33:32.209 "name": "BaseBdev2", 00:33:32.209 "aliases": [ 00:33:32.209 "6a55d2ba-9313-4ed8-9387-796ea8c4768f" 00:33:32.209 ], 00:33:32.209 "product_name": "Malloc disk", 00:33:32.209 "block_size": 512, 00:33:32.209 "num_blocks": 65536, 00:33:32.209 "uuid": "6a55d2ba-9313-4ed8-9387-796ea8c4768f", 00:33:32.209 "assigned_rate_limits": { 00:33:32.209 "rw_ios_per_sec": 0, 00:33:32.209 "rw_mbytes_per_sec": 0, 00:33:32.209 "r_mbytes_per_sec": 0, 00:33:32.209 "w_mbytes_per_sec": 0 00:33:32.209 }, 00:33:32.209 "claimed": false, 00:33:32.209 "zoned": false, 00:33:32.209 "supported_io_types": { 00:33:32.209 "read": true, 00:33:32.209 "write": true, 00:33:32.209 "unmap": true, 00:33:32.209 "flush": true, 00:33:32.209 "reset": true, 00:33:32.209 "nvme_admin": false, 00:33:32.209 "nvme_io": false, 00:33:32.209 "nvme_io_md": false, 00:33:32.209 "write_zeroes": true, 00:33:32.209 "zcopy": true, 00:33:32.209 "get_zone_info": false, 00:33:32.209 "zone_management": false, 00:33:32.209 "zone_append": false, 00:33:32.209 "compare": false, 00:33:32.209 "compare_and_write": false, 00:33:32.209 "abort": true, 00:33:32.209 "seek_hole": false, 00:33:32.209 "seek_data": false, 00:33:32.209 "copy": true, 00:33:32.209 "nvme_iov_md": false 00:33:32.209 }, 00:33:32.209 "memory_domains": [ 00:33:32.209 { 00:33:32.209 "dma_device_id": "system", 00:33:32.209 "dma_device_type": 1 00:33:32.209 }, 00:33:32.209 { 00:33:32.209 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:32.209 "dma_device_type": 2 00:33:32.209 } 00:33:32.209 ], 00:33:32.209 "driver_specific": {} 00:33:32.209 } 00:33:32.209 ] 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:32.209 BaseBdev3 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:32.209 [ 00:33:32.209 { 00:33:32.209 "name": "BaseBdev3", 00:33:32.209 "aliases": [ 00:33:32.209 "4f67f0bd-f547-467d-93d1-edf223eb909d" 00:33:32.209 ], 00:33:32.209 "product_name": "Malloc disk", 00:33:32.209 "block_size": 512, 00:33:32.209 "num_blocks": 65536, 00:33:32.209 "uuid": "4f67f0bd-f547-467d-93d1-edf223eb909d", 00:33:32.209 "assigned_rate_limits": { 00:33:32.209 "rw_ios_per_sec": 0, 00:33:32.209 "rw_mbytes_per_sec": 0, 00:33:32.209 "r_mbytes_per_sec": 0, 00:33:32.209 "w_mbytes_per_sec": 0 00:33:32.209 }, 00:33:32.209 "claimed": false, 00:33:32.209 "zoned": false, 00:33:32.209 "supported_io_types": { 00:33:32.209 "read": true, 00:33:32.209 "write": true, 00:33:32.209 "unmap": true, 00:33:32.209 "flush": true, 00:33:32.209 "reset": true, 00:33:32.209 "nvme_admin": false, 00:33:32.209 "nvme_io": false, 00:33:32.209 "nvme_io_md": false, 00:33:32.209 "write_zeroes": true, 00:33:32.209 "zcopy": true, 00:33:32.209 "get_zone_info": false, 00:33:32.209 "zone_management": false, 00:33:32.209 "zone_append": false, 00:33:32.209 "compare": false, 00:33:32.209 "compare_and_write": false, 00:33:32.209 "abort": true, 00:33:32.209 "seek_hole": false, 00:33:32.209 "seek_data": false, 00:33:32.209 "copy": true, 00:33:32.209 "nvme_iov_md": false 00:33:32.209 }, 00:33:32.209 "memory_domains": [ 00:33:32.209 { 00:33:32.209 "dma_device_id": "system", 00:33:32.209 "dma_device_type": 1 00:33:32.209 }, 00:33:32.209 { 00:33:32.209 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:32.209 "dma_device_type": 2 00:33:32.209 } 00:33:32.209 ], 00:33:32.209 "driver_specific": {} 00:33:32.209 } 00:33:32.209 ] 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:32.209 BaseBdev4 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:32.209 [ 00:33:32.209 { 00:33:32.209 "name": "BaseBdev4", 00:33:32.209 "aliases": [ 00:33:32.209 "c902db70-8a5f-41f9-9f39-d864f848b51d" 00:33:32.209 ], 00:33:32.209 "product_name": "Malloc disk", 00:33:32.209 "block_size": 512, 00:33:32.209 "num_blocks": 65536, 00:33:32.209 "uuid": "c902db70-8a5f-41f9-9f39-d864f848b51d", 00:33:32.209 "assigned_rate_limits": { 00:33:32.209 "rw_ios_per_sec": 0, 00:33:32.209 "rw_mbytes_per_sec": 0, 00:33:32.209 "r_mbytes_per_sec": 0, 00:33:32.209 "w_mbytes_per_sec": 0 00:33:32.209 }, 00:33:32.209 "claimed": false, 00:33:32.209 "zoned": false, 00:33:32.209 "supported_io_types": { 00:33:32.209 "read": true, 00:33:32.209 "write": true, 00:33:32.209 "unmap": true, 00:33:32.209 "flush": true, 00:33:32.209 "reset": true, 00:33:32.209 "nvme_admin": false, 00:33:32.209 "nvme_io": false, 00:33:32.209 "nvme_io_md": false, 00:33:32.209 "write_zeroes": true, 00:33:32.209 "zcopy": true, 00:33:32.209 "get_zone_info": false, 00:33:32.209 "zone_management": false, 00:33:32.209 "zone_append": false, 00:33:32.209 "compare": false, 00:33:32.209 "compare_and_write": false, 00:33:32.209 "abort": true, 00:33:32.209 "seek_hole": false, 00:33:32.209 "seek_data": false, 00:33:32.209 "copy": true, 00:33:32.209 "nvme_iov_md": false 00:33:32.209 }, 00:33:32.209 "memory_domains": [ 00:33:32.209 { 00:33:32.209 "dma_device_id": "system", 00:33:32.209 "dma_device_type": 1 00:33:32.209 }, 00:33:32.209 { 00:33:32.209 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:32.209 "dma_device_type": 2 00:33:32.209 } 00:33:32.209 ], 00:33:32.209 "driver_specific": {} 00:33:32.209 } 00:33:32.209 ] 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:33:32.209 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:33:32.210 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:33:32.210 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:32.210 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:32.210 [2024-11-20 13:53:39.254928] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:33:32.210 [2024-11-20 13:53:39.255107] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:33:32.210 [2024-11-20 13:53:39.255182] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:33:32.210 [2024-11-20 13:53:39.257168] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:33:32.210 [2024-11-20 13:53:39.257297] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:33:32.210 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:32.210 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:33:32.210 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:32.210 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:32.210 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:32.210 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:32.210 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:32.210 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:32.210 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:32.210 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:32.210 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:32.210 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:32.210 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:32.210 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:32.210 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:32.469 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:32.469 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:32.469 "name": "Existed_Raid", 00:33:32.469 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:32.469 "strip_size_kb": 64, 00:33:32.469 "state": "configuring", 00:33:32.469 "raid_level": "raid5f", 00:33:32.469 "superblock": false, 00:33:32.469 "num_base_bdevs": 4, 00:33:32.469 "num_base_bdevs_discovered": 3, 00:33:32.469 "num_base_bdevs_operational": 4, 00:33:32.469 "base_bdevs_list": [ 00:33:32.469 { 00:33:32.469 "name": "BaseBdev1", 00:33:32.469 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:32.469 "is_configured": false, 00:33:32.469 "data_offset": 0, 00:33:32.469 "data_size": 0 00:33:32.469 }, 00:33:32.469 { 00:33:32.469 "name": "BaseBdev2", 00:33:32.469 "uuid": "6a55d2ba-9313-4ed8-9387-796ea8c4768f", 00:33:32.469 "is_configured": true, 00:33:32.469 "data_offset": 0, 00:33:32.469 "data_size": 65536 00:33:32.469 }, 00:33:32.469 { 00:33:32.469 "name": "BaseBdev3", 00:33:32.469 "uuid": "4f67f0bd-f547-467d-93d1-edf223eb909d", 00:33:32.469 "is_configured": true, 00:33:32.469 "data_offset": 0, 00:33:32.469 "data_size": 65536 00:33:32.469 }, 00:33:32.469 { 00:33:32.469 "name": "BaseBdev4", 00:33:32.469 "uuid": "c902db70-8a5f-41f9-9f39-d864f848b51d", 00:33:32.469 "is_configured": true, 00:33:32.469 "data_offset": 0, 00:33:32.469 "data_size": 65536 00:33:32.469 } 00:33:32.469 ] 00:33:32.469 }' 00:33:32.469 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:32.469 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:32.727 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:33:32.728 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:32.728 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:32.728 [2024-11-20 13:53:39.599049] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:33:32.728 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:32.728 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:33:32.728 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:32.728 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:32.728 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:32.728 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:32.728 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:32.728 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:32.728 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:32.728 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:32.728 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:32.728 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:32.728 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:32.728 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:32.728 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:32.728 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:32.728 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:32.728 "name": "Existed_Raid", 00:33:32.728 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:32.728 "strip_size_kb": 64, 00:33:32.728 "state": "configuring", 00:33:32.728 "raid_level": "raid5f", 00:33:32.728 "superblock": false, 00:33:32.728 "num_base_bdevs": 4, 00:33:32.728 "num_base_bdevs_discovered": 2, 00:33:32.728 "num_base_bdevs_operational": 4, 00:33:32.728 "base_bdevs_list": [ 00:33:32.728 { 00:33:32.728 "name": "BaseBdev1", 00:33:32.728 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:32.728 "is_configured": false, 00:33:32.728 "data_offset": 0, 00:33:32.728 "data_size": 0 00:33:32.728 }, 00:33:32.728 { 00:33:32.728 "name": null, 00:33:32.728 "uuid": "6a55d2ba-9313-4ed8-9387-796ea8c4768f", 00:33:32.728 "is_configured": false, 00:33:32.728 "data_offset": 0, 00:33:32.728 "data_size": 65536 00:33:32.728 }, 00:33:32.728 { 00:33:32.728 "name": "BaseBdev3", 00:33:32.728 "uuid": "4f67f0bd-f547-467d-93d1-edf223eb909d", 00:33:32.728 "is_configured": true, 00:33:32.728 "data_offset": 0, 00:33:32.728 "data_size": 65536 00:33:32.728 }, 00:33:32.728 { 00:33:32.728 "name": "BaseBdev4", 00:33:32.728 "uuid": "c902db70-8a5f-41f9-9f39-d864f848b51d", 00:33:32.728 "is_configured": true, 00:33:32.728 "data_offset": 0, 00:33:32.728 "data_size": 65536 00:33:32.728 } 00:33:32.728 ] 00:33:32.728 }' 00:33:32.728 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:32.728 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:32.986 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:32.986 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:32.986 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:32.986 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:33:32.986 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:32.987 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:33:32.987 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:33:32.987 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:32.987 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:32.987 [2024-11-20 13:53:39.991828] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:33:32.987 BaseBdev1 00:33:32.987 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:32.987 13:53:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:33:32.987 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:33:32.987 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:33:32.987 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:33:32.987 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:33:32.987 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:33:32.987 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:33:32.987 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:32.987 13:53:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:32.987 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:32.987 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:33:32.987 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:32.987 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:32.987 [ 00:33:32.987 { 00:33:32.987 "name": "BaseBdev1", 00:33:32.987 "aliases": [ 00:33:32.987 "49ad7886-866e-4f8c-8a04-33881ddfe494" 00:33:32.987 ], 00:33:32.987 "product_name": "Malloc disk", 00:33:32.987 "block_size": 512, 00:33:32.987 "num_blocks": 65536, 00:33:32.987 "uuid": "49ad7886-866e-4f8c-8a04-33881ddfe494", 00:33:32.987 "assigned_rate_limits": { 00:33:32.987 "rw_ios_per_sec": 0, 00:33:32.987 "rw_mbytes_per_sec": 0, 00:33:32.987 "r_mbytes_per_sec": 0, 00:33:32.987 "w_mbytes_per_sec": 0 00:33:32.987 }, 00:33:32.987 "claimed": true, 00:33:32.987 "claim_type": "exclusive_write", 00:33:32.987 "zoned": false, 00:33:32.987 "supported_io_types": { 00:33:32.987 "read": true, 00:33:32.987 "write": true, 00:33:32.987 "unmap": true, 00:33:32.987 "flush": true, 00:33:32.987 "reset": true, 00:33:32.987 "nvme_admin": false, 00:33:32.987 "nvme_io": false, 00:33:32.987 "nvme_io_md": false, 00:33:32.987 "write_zeroes": true, 00:33:32.987 "zcopy": true, 00:33:32.987 "get_zone_info": false, 00:33:32.987 "zone_management": false, 00:33:32.987 "zone_append": false, 00:33:32.987 "compare": false, 00:33:32.987 "compare_and_write": false, 00:33:32.987 "abort": true, 00:33:32.987 "seek_hole": false, 00:33:32.987 "seek_data": false, 00:33:32.987 "copy": true, 00:33:32.987 "nvme_iov_md": false 00:33:32.987 }, 00:33:32.987 "memory_domains": [ 00:33:32.987 { 00:33:32.987 "dma_device_id": "system", 00:33:32.987 "dma_device_type": 1 00:33:32.987 }, 00:33:32.987 { 00:33:32.987 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:32.987 "dma_device_type": 2 00:33:32.987 } 00:33:32.987 ], 00:33:32.987 "driver_specific": {} 00:33:32.987 } 00:33:32.987 ] 00:33:32.987 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:32.987 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:33:32.987 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:33:32.987 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:32.987 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:32.987 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:32.987 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:32.987 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:32.987 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:32.987 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:32.987 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:32.987 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:32.987 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:32.987 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:32.987 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:32.987 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:32.987 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:33.245 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:33.245 "name": "Existed_Raid", 00:33:33.245 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:33.245 "strip_size_kb": 64, 00:33:33.245 "state": "configuring", 00:33:33.245 "raid_level": "raid5f", 00:33:33.246 "superblock": false, 00:33:33.246 "num_base_bdevs": 4, 00:33:33.246 "num_base_bdevs_discovered": 3, 00:33:33.246 "num_base_bdevs_operational": 4, 00:33:33.246 "base_bdevs_list": [ 00:33:33.246 { 00:33:33.246 "name": "BaseBdev1", 00:33:33.246 "uuid": "49ad7886-866e-4f8c-8a04-33881ddfe494", 00:33:33.246 "is_configured": true, 00:33:33.246 "data_offset": 0, 00:33:33.246 "data_size": 65536 00:33:33.246 }, 00:33:33.246 { 00:33:33.246 "name": null, 00:33:33.246 "uuid": "6a55d2ba-9313-4ed8-9387-796ea8c4768f", 00:33:33.246 "is_configured": false, 00:33:33.246 "data_offset": 0, 00:33:33.246 "data_size": 65536 00:33:33.246 }, 00:33:33.246 { 00:33:33.246 "name": "BaseBdev3", 00:33:33.246 "uuid": "4f67f0bd-f547-467d-93d1-edf223eb909d", 00:33:33.246 "is_configured": true, 00:33:33.246 "data_offset": 0, 00:33:33.246 "data_size": 65536 00:33:33.246 }, 00:33:33.246 { 00:33:33.246 "name": "BaseBdev4", 00:33:33.246 "uuid": "c902db70-8a5f-41f9-9f39-d864f848b51d", 00:33:33.246 "is_configured": true, 00:33:33.246 "data_offset": 0, 00:33:33.246 "data_size": 65536 00:33:33.246 } 00:33:33.246 ] 00:33:33.246 }' 00:33:33.246 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:33.246 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:33.505 [2024-11-20 13:53:40.371990] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:33.505 "name": "Existed_Raid", 00:33:33.505 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:33.505 "strip_size_kb": 64, 00:33:33.505 "state": "configuring", 00:33:33.505 "raid_level": "raid5f", 00:33:33.505 "superblock": false, 00:33:33.505 "num_base_bdevs": 4, 00:33:33.505 "num_base_bdevs_discovered": 2, 00:33:33.505 "num_base_bdevs_operational": 4, 00:33:33.505 "base_bdevs_list": [ 00:33:33.505 { 00:33:33.505 "name": "BaseBdev1", 00:33:33.505 "uuid": "49ad7886-866e-4f8c-8a04-33881ddfe494", 00:33:33.505 "is_configured": true, 00:33:33.505 "data_offset": 0, 00:33:33.505 "data_size": 65536 00:33:33.505 }, 00:33:33.505 { 00:33:33.505 "name": null, 00:33:33.505 "uuid": "6a55d2ba-9313-4ed8-9387-796ea8c4768f", 00:33:33.505 "is_configured": false, 00:33:33.505 "data_offset": 0, 00:33:33.505 "data_size": 65536 00:33:33.505 }, 00:33:33.505 { 00:33:33.505 "name": null, 00:33:33.505 "uuid": "4f67f0bd-f547-467d-93d1-edf223eb909d", 00:33:33.505 "is_configured": false, 00:33:33.505 "data_offset": 0, 00:33:33.505 "data_size": 65536 00:33:33.505 }, 00:33:33.505 { 00:33:33.505 "name": "BaseBdev4", 00:33:33.505 "uuid": "c902db70-8a5f-41f9-9f39-d864f848b51d", 00:33:33.505 "is_configured": true, 00:33:33.505 "data_offset": 0, 00:33:33.505 "data_size": 65536 00:33:33.505 } 00:33:33.505 ] 00:33:33.505 }' 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:33.505 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:33.764 [2024-11-20 13:53:40.772065] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:33.764 "name": "Existed_Raid", 00:33:33.764 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:33.764 "strip_size_kb": 64, 00:33:33.764 "state": "configuring", 00:33:33.764 "raid_level": "raid5f", 00:33:33.764 "superblock": false, 00:33:33.764 "num_base_bdevs": 4, 00:33:33.764 "num_base_bdevs_discovered": 3, 00:33:33.764 "num_base_bdevs_operational": 4, 00:33:33.764 "base_bdevs_list": [ 00:33:33.764 { 00:33:33.764 "name": "BaseBdev1", 00:33:33.764 "uuid": "49ad7886-866e-4f8c-8a04-33881ddfe494", 00:33:33.764 "is_configured": true, 00:33:33.764 "data_offset": 0, 00:33:33.764 "data_size": 65536 00:33:33.764 }, 00:33:33.764 { 00:33:33.764 "name": null, 00:33:33.764 "uuid": "6a55d2ba-9313-4ed8-9387-796ea8c4768f", 00:33:33.764 "is_configured": false, 00:33:33.764 "data_offset": 0, 00:33:33.764 "data_size": 65536 00:33:33.764 }, 00:33:33.764 { 00:33:33.764 "name": "BaseBdev3", 00:33:33.764 "uuid": "4f67f0bd-f547-467d-93d1-edf223eb909d", 00:33:33.764 "is_configured": true, 00:33:33.764 "data_offset": 0, 00:33:33.764 "data_size": 65536 00:33:33.764 }, 00:33:33.764 { 00:33:33.764 "name": "BaseBdev4", 00:33:33.764 "uuid": "c902db70-8a5f-41f9-9f39-d864f848b51d", 00:33:33.764 "is_configured": true, 00:33:33.764 "data_offset": 0, 00:33:33.764 "data_size": 65536 00:33:33.764 } 00:33:33.764 ] 00:33:33.764 }' 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:33.764 13:53:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:34.330 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:33:34.330 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:34.330 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:34.330 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:34.330 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:34.330 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:33:34.330 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:33:34.330 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:34.330 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:34.330 [2024-11-20 13:53:41.128143] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:33:34.330 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:34.330 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:33:34.330 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:34.330 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:34.330 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:34.330 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:34.330 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:34.330 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:34.330 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:34.330 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:34.330 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:34.330 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:34.330 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:34.330 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:34.330 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:34.330 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:34.330 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:34.330 "name": "Existed_Raid", 00:33:34.331 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:34.331 "strip_size_kb": 64, 00:33:34.331 "state": "configuring", 00:33:34.331 "raid_level": "raid5f", 00:33:34.331 "superblock": false, 00:33:34.331 "num_base_bdevs": 4, 00:33:34.331 "num_base_bdevs_discovered": 2, 00:33:34.331 "num_base_bdevs_operational": 4, 00:33:34.331 "base_bdevs_list": [ 00:33:34.331 { 00:33:34.331 "name": null, 00:33:34.331 "uuid": "49ad7886-866e-4f8c-8a04-33881ddfe494", 00:33:34.331 "is_configured": false, 00:33:34.331 "data_offset": 0, 00:33:34.331 "data_size": 65536 00:33:34.331 }, 00:33:34.331 { 00:33:34.331 "name": null, 00:33:34.331 "uuid": "6a55d2ba-9313-4ed8-9387-796ea8c4768f", 00:33:34.331 "is_configured": false, 00:33:34.331 "data_offset": 0, 00:33:34.331 "data_size": 65536 00:33:34.331 }, 00:33:34.331 { 00:33:34.331 "name": "BaseBdev3", 00:33:34.331 "uuid": "4f67f0bd-f547-467d-93d1-edf223eb909d", 00:33:34.331 "is_configured": true, 00:33:34.331 "data_offset": 0, 00:33:34.331 "data_size": 65536 00:33:34.331 }, 00:33:34.331 { 00:33:34.331 "name": "BaseBdev4", 00:33:34.331 "uuid": "c902db70-8a5f-41f9-9f39-d864f848b51d", 00:33:34.331 "is_configured": true, 00:33:34.331 "data_offset": 0, 00:33:34.331 "data_size": 65536 00:33:34.331 } 00:33:34.331 ] 00:33:34.331 }' 00:33:34.331 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:34.331 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:34.589 [2024-11-20 13:53:41.537278] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:34.589 "name": "Existed_Raid", 00:33:34.589 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:34.589 "strip_size_kb": 64, 00:33:34.589 "state": "configuring", 00:33:34.589 "raid_level": "raid5f", 00:33:34.589 "superblock": false, 00:33:34.589 "num_base_bdevs": 4, 00:33:34.589 "num_base_bdevs_discovered": 3, 00:33:34.589 "num_base_bdevs_operational": 4, 00:33:34.589 "base_bdevs_list": [ 00:33:34.589 { 00:33:34.589 "name": null, 00:33:34.589 "uuid": "49ad7886-866e-4f8c-8a04-33881ddfe494", 00:33:34.589 "is_configured": false, 00:33:34.589 "data_offset": 0, 00:33:34.589 "data_size": 65536 00:33:34.589 }, 00:33:34.589 { 00:33:34.589 "name": "BaseBdev2", 00:33:34.589 "uuid": "6a55d2ba-9313-4ed8-9387-796ea8c4768f", 00:33:34.589 "is_configured": true, 00:33:34.589 "data_offset": 0, 00:33:34.589 "data_size": 65536 00:33:34.589 }, 00:33:34.589 { 00:33:34.589 "name": "BaseBdev3", 00:33:34.589 "uuid": "4f67f0bd-f547-467d-93d1-edf223eb909d", 00:33:34.589 "is_configured": true, 00:33:34.589 "data_offset": 0, 00:33:34.589 "data_size": 65536 00:33:34.589 }, 00:33:34.589 { 00:33:34.589 "name": "BaseBdev4", 00:33:34.589 "uuid": "c902db70-8a5f-41f9-9f39-d864f848b51d", 00:33:34.589 "is_configured": true, 00:33:34.589 "data_offset": 0, 00:33:34.589 "data_size": 65536 00:33:34.589 } 00:33:34.589 ] 00:33:34.589 }' 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:34.589 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:34.847 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:33:34.847 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:34.848 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:34.848 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:34.848 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:34.848 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:33:34.848 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:34.848 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:33:34.848 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:34.848 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:34.848 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 49ad7886-866e-4f8c-8a04-33881ddfe494 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:35.106 [2024-11-20 13:53:41.929658] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:33:35.106 [2024-11-20 13:53:41.929820] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:33:35.106 [2024-11-20 13:53:41.929833] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:33:35.106 [2024-11-20 13:53:41.930106] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:33:35.106 [2024-11-20 13:53:41.934091] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:33:35.106 [2024-11-20 13:53:41.934169] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:33:35.106 [2024-11-20 13:53:41.934449] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:35.106 NewBaseBdev 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:35.106 [ 00:33:35.106 { 00:33:35.106 "name": "NewBaseBdev", 00:33:35.106 "aliases": [ 00:33:35.106 "49ad7886-866e-4f8c-8a04-33881ddfe494" 00:33:35.106 ], 00:33:35.106 "product_name": "Malloc disk", 00:33:35.106 "block_size": 512, 00:33:35.106 "num_blocks": 65536, 00:33:35.106 "uuid": "49ad7886-866e-4f8c-8a04-33881ddfe494", 00:33:35.106 "assigned_rate_limits": { 00:33:35.106 "rw_ios_per_sec": 0, 00:33:35.106 "rw_mbytes_per_sec": 0, 00:33:35.106 "r_mbytes_per_sec": 0, 00:33:35.106 "w_mbytes_per_sec": 0 00:33:35.106 }, 00:33:35.106 "claimed": true, 00:33:35.106 "claim_type": "exclusive_write", 00:33:35.106 "zoned": false, 00:33:35.106 "supported_io_types": { 00:33:35.106 "read": true, 00:33:35.106 "write": true, 00:33:35.106 "unmap": true, 00:33:35.106 "flush": true, 00:33:35.106 "reset": true, 00:33:35.106 "nvme_admin": false, 00:33:35.106 "nvme_io": false, 00:33:35.106 "nvme_io_md": false, 00:33:35.106 "write_zeroes": true, 00:33:35.106 "zcopy": true, 00:33:35.106 "get_zone_info": false, 00:33:35.106 "zone_management": false, 00:33:35.106 "zone_append": false, 00:33:35.106 "compare": false, 00:33:35.106 "compare_and_write": false, 00:33:35.106 "abort": true, 00:33:35.106 "seek_hole": false, 00:33:35.106 "seek_data": false, 00:33:35.106 "copy": true, 00:33:35.106 "nvme_iov_md": false 00:33:35.106 }, 00:33:35.106 "memory_domains": [ 00:33:35.106 { 00:33:35.106 "dma_device_id": "system", 00:33:35.106 "dma_device_type": 1 00:33:35.106 }, 00:33:35.106 { 00:33:35.106 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:35.106 "dma_device_type": 2 00:33:35.106 } 00:33:35.106 ], 00:33:35.106 "driver_specific": {} 00:33:35.106 } 00:33:35.106 ] 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:35.106 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:35.106 "name": "Existed_Raid", 00:33:35.106 "uuid": "5099adbe-ca84-4838-8a92-c15e2942587a", 00:33:35.106 "strip_size_kb": 64, 00:33:35.106 "state": "online", 00:33:35.106 "raid_level": "raid5f", 00:33:35.106 "superblock": false, 00:33:35.106 "num_base_bdevs": 4, 00:33:35.106 "num_base_bdevs_discovered": 4, 00:33:35.106 "num_base_bdevs_operational": 4, 00:33:35.106 "base_bdevs_list": [ 00:33:35.106 { 00:33:35.106 "name": "NewBaseBdev", 00:33:35.106 "uuid": "49ad7886-866e-4f8c-8a04-33881ddfe494", 00:33:35.106 "is_configured": true, 00:33:35.106 "data_offset": 0, 00:33:35.106 "data_size": 65536 00:33:35.106 }, 00:33:35.106 { 00:33:35.106 "name": "BaseBdev2", 00:33:35.106 "uuid": "6a55d2ba-9313-4ed8-9387-796ea8c4768f", 00:33:35.107 "is_configured": true, 00:33:35.107 "data_offset": 0, 00:33:35.107 "data_size": 65536 00:33:35.107 }, 00:33:35.107 { 00:33:35.107 "name": "BaseBdev3", 00:33:35.107 "uuid": "4f67f0bd-f547-467d-93d1-edf223eb909d", 00:33:35.107 "is_configured": true, 00:33:35.107 "data_offset": 0, 00:33:35.107 "data_size": 65536 00:33:35.107 }, 00:33:35.107 { 00:33:35.107 "name": "BaseBdev4", 00:33:35.107 "uuid": "c902db70-8a5f-41f9-9f39-d864f848b51d", 00:33:35.107 "is_configured": true, 00:33:35.107 "data_offset": 0, 00:33:35.107 "data_size": 65536 00:33:35.107 } 00:33:35.107 ] 00:33:35.107 }' 00:33:35.107 13:53:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:35.107 13:53:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:35.365 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:33:35.365 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:33:35.365 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:33:35.365 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:33:35.365 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:33:35.365 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:33:35.365 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:33:35.365 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:35.365 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:33:35.365 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:35.365 [2024-11-20 13:53:42.275378] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:33:35.365 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:35.365 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:33:35.365 "name": "Existed_Raid", 00:33:35.365 "aliases": [ 00:33:35.365 "5099adbe-ca84-4838-8a92-c15e2942587a" 00:33:35.365 ], 00:33:35.365 "product_name": "Raid Volume", 00:33:35.365 "block_size": 512, 00:33:35.365 "num_blocks": 196608, 00:33:35.365 "uuid": "5099adbe-ca84-4838-8a92-c15e2942587a", 00:33:35.365 "assigned_rate_limits": { 00:33:35.365 "rw_ios_per_sec": 0, 00:33:35.365 "rw_mbytes_per_sec": 0, 00:33:35.365 "r_mbytes_per_sec": 0, 00:33:35.365 "w_mbytes_per_sec": 0 00:33:35.365 }, 00:33:35.365 "claimed": false, 00:33:35.365 "zoned": false, 00:33:35.365 "supported_io_types": { 00:33:35.365 "read": true, 00:33:35.365 "write": true, 00:33:35.365 "unmap": false, 00:33:35.365 "flush": false, 00:33:35.365 "reset": true, 00:33:35.365 "nvme_admin": false, 00:33:35.365 "nvme_io": false, 00:33:35.365 "nvme_io_md": false, 00:33:35.365 "write_zeroes": true, 00:33:35.365 "zcopy": false, 00:33:35.365 "get_zone_info": false, 00:33:35.365 "zone_management": false, 00:33:35.365 "zone_append": false, 00:33:35.365 "compare": false, 00:33:35.365 "compare_and_write": false, 00:33:35.365 "abort": false, 00:33:35.365 "seek_hole": false, 00:33:35.365 "seek_data": false, 00:33:35.365 "copy": false, 00:33:35.365 "nvme_iov_md": false 00:33:35.365 }, 00:33:35.365 "driver_specific": { 00:33:35.365 "raid": { 00:33:35.365 "uuid": "5099adbe-ca84-4838-8a92-c15e2942587a", 00:33:35.365 "strip_size_kb": 64, 00:33:35.365 "state": "online", 00:33:35.365 "raid_level": "raid5f", 00:33:35.365 "superblock": false, 00:33:35.365 "num_base_bdevs": 4, 00:33:35.365 "num_base_bdevs_discovered": 4, 00:33:35.365 "num_base_bdevs_operational": 4, 00:33:35.365 "base_bdevs_list": [ 00:33:35.365 { 00:33:35.365 "name": "NewBaseBdev", 00:33:35.365 "uuid": "49ad7886-866e-4f8c-8a04-33881ddfe494", 00:33:35.365 "is_configured": true, 00:33:35.365 "data_offset": 0, 00:33:35.365 "data_size": 65536 00:33:35.365 }, 00:33:35.365 { 00:33:35.365 "name": "BaseBdev2", 00:33:35.365 "uuid": "6a55d2ba-9313-4ed8-9387-796ea8c4768f", 00:33:35.365 "is_configured": true, 00:33:35.365 "data_offset": 0, 00:33:35.365 "data_size": 65536 00:33:35.365 }, 00:33:35.365 { 00:33:35.365 "name": "BaseBdev3", 00:33:35.365 "uuid": "4f67f0bd-f547-467d-93d1-edf223eb909d", 00:33:35.365 "is_configured": true, 00:33:35.365 "data_offset": 0, 00:33:35.365 "data_size": 65536 00:33:35.365 }, 00:33:35.365 { 00:33:35.365 "name": "BaseBdev4", 00:33:35.365 "uuid": "c902db70-8a5f-41f9-9f39-d864f848b51d", 00:33:35.365 "is_configured": true, 00:33:35.365 "data_offset": 0, 00:33:35.365 "data_size": 65536 00:33:35.365 } 00:33:35.365 ] 00:33:35.365 } 00:33:35.365 } 00:33:35.365 }' 00:33:35.365 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:33:35.365 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:33:35.365 BaseBdev2 00:33:35.365 BaseBdev3 00:33:35.365 BaseBdev4' 00:33:35.365 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:35.365 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:33:35.365 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:35.365 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:33:35.365 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:35.365 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:35.366 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:35.366 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:35.366 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:33:35.366 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:33:35.366 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:35.366 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:33:35.366 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:35.366 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:35.366 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:35.366 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:35.624 [2024-11-20 13:53:42.499188] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:33:35.624 [2024-11-20 13:53:42.499217] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:33:35.624 [2024-11-20 13:53:42.499289] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:33:35.624 [2024-11-20 13:53:42.499555] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:33:35.624 [2024-11-20 13:53:42.499564] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 80385 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 80385 ']' 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@958 -- # kill -0 80385 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # uname 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 80385 00:33:35.624 killing process with pid 80385 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 80385' 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@973 -- # kill 80385 00:33:35.624 [2024-11-20 13:53:42.531939] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:33:35.624 13:53:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@978 -- # wait 80385 00:33:35.882 [2024-11-20 13:53:42.736320] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:33:36.446 ************************************ 00:33:36.446 END TEST raid5f_state_function_test 00:33:36.446 ************************************ 00:33:36.446 13:53:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:33:36.446 00:33:36.446 real 0m8.358s 00:33:36.446 user 0m13.342s 00:33:36.446 sys 0m1.422s 00:33:36.446 13:53:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:33:36.446 13:53:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:33:36.446 13:53:43 bdev_raid -- bdev/bdev_raid.sh@987 -- # run_test raid5f_state_function_test_sb raid_state_function_test raid5f 4 true 00:33:36.446 13:53:43 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:33:36.446 13:53:43 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:33:36.446 13:53:43 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:33:36.446 ************************************ 00:33:36.446 START TEST raid5f_state_function_test_sb 00:33:36.446 ************************************ 00:33:36.446 13:53:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 4 true 00:33:36.446 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:33:36.446 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:33:36.446 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:33:36.446 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:33:36.446 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:33:36.446 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:33:36.446 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:33:36.446 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:33:36.446 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:33:36.446 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:33:36.446 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:33:36.446 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:33:36.446 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:33:36.446 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:33:36.446 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:33:36.446 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:33:36.446 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:33:36.446 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:33:36.446 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:33:36.446 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:33:36.446 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:33:36.446 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:33:36.447 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:33:36.447 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:33:36.447 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:33:36.447 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:33:36.447 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:33:36.447 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:33:36.447 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:33:36.447 Process raid pid: 81018 00:33:36.447 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=81018 00:33:36.447 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:33:36.447 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 81018' 00:33:36.447 13:53:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 81018 00:33:36.447 13:53:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 81018 ']' 00:33:36.447 13:53:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:33:36.447 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:33:36.447 13:53:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:33:36.447 13:53:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:33:36.447 13:53:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:33:36.447 13:53:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:36.447 [2024-11-20 13:53:43.497948] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:33:36.447 [2024-11-20 13:53:43.498094] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:33:36.705 [2024-11-20 13:53:43.662595] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:33:36.962 [2024-11-20 13:53:43.781808] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:33:36.962 [2024-11-20 13:53:43.940379] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:33:36.962 [2024-11-20 13:53:43.940615] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:33:37.529 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:33:37.529 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:33:37.529 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:33:37.529 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:37.529 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:37.529 [2024-11-20 13:53:44.330366] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:33:37.529 [2024-11-20 13:53:44.330426] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:33:37.529 [2024-11-20 13:53:44.330437] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:33:37.529 [2024-11-20 13:53:44.330447] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:33:37.529 [2024-11-20 13:53:44.330454] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:33:37.529 [2024-11-20 13:53:44.330463] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:33:37.529 [2024-11-20 13:53:44.330469] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:33:37.529 [2024-11-20 13:53:44.330478] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:33:37.529 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:37.529 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:33:37.529 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:37.529 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:37.529 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:37.529 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:37.529 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:37.529 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:37.529 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:37.529 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:37.529 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:37.529 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:37.529 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:37.529 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:37.529 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:37.529 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:37.529 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:37.529 "name": "Existed_Raid", 00:33:37.529 "uuid": "2e9858ef-cd97-4b04-a19c-13ca41c2b967", 00:33:37.529 "strip_size_kb": 64, 00:33:37.529 "state": "configuring", 00:33:37.529 "raid_level": "raid5f", 00:33:37.529 "superblock": true, 00:33:37.529 "num_base_bdevs": 4, 00:33:37.529 "num_base_bdevs_discovered": 0, 00:33:37.529 "num_base_bdevs_operational": 4, 00:33:37.529 "base_bdevs_list": [ 00:33:37.529 { 00:33:37.529 "name": "BaseBdev1", 00:33:37.529 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:37.529 "is_configured": false, 00:33:37.529 "data_offset": 0, 00:33:37.529 "data_size": 0 00:33:37.529 }, 00:33:37.529 { 00:33:37.529 "name": "BaseBdev2", 00:33:37.529 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:37.529 "is_configured": false, 00:33:37.529 "data_offset": 0, 00:33:37.529 "data_size": 0 00:33:37.529 }, 00:33:37.529 { 00:33:37.529 "name": "BaseBdev3", 00:33:37.529 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:37.529 "is_configured": false, 00:33:37.529 "data_offset": 0, 00:33:37.529 "data_size": 0 00:33:37.529 }, 00:33:37.529 { 00:33:37.529 "name": "BaseBdev4", 00:33:37.529 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:37.529 "is_configured": false, 00:33:37.529 "data_offset": 0, 00:33:37.529 "data_size": 0 00:33:37.529 } 00:33:37.529 ] 00:33:37.529 }' 00:33:37.529 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:37.529 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:37.788 [2024-11-20 13:53:44.654358] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:33:37.788 [2024-11-20 13:53:44.654397] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:37.788 [2024-11-20 13:53:44.662359] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:33:37.788 [2024-11-20 13:53:44.662398] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:33:37.788 [2024-11-20 13:53:44.662406] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:33:37.788 [2024-11-20 13:53:44.662416] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:33:37.788 [2024-11-20 13:53:44.662422] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:33:37.788 [2024-11-20 13:53:44.662431] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:33:37.788 [2024-11-20 13:53:44.662437] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:33:37.788 [2024-11-20 13:53:44.662445] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:37.788 [2024-11-20 13:53:44.697190] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:33:37.788 BaseBdev1 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:37.788 [ 00:33:37.788 { 00:33:37.788 "name": "BaseBdev1", 00:33:37.788 "aliases": [ 00:33:37.788 "0e46dddb-e2f4-4706-9dd8-026901d0168f" 00:33:37.788 ], 00:33:37.788 "product_name": "Malloc disk", 00:33:37.788 "block_size": 512, 00:33:37.788 "num_blocks": 65536, 00:33:37.788 "uuid": "0e46dddb-e2f4-4706-9dd8-026901d0168f", 00:33:37.788 "assigned_rate_limits": { 00:33:37.788 "rw_ios_per_sec": 0, 00:33:37.788 "rw_mbytes_per_sec": 0, 00:33:37.788 "r_mbytes_per_sec": 0, 00:33:37.788 "w_mbytes_per_sec": 0 00:33:37.788 }, 00:33:37.788 "claimed": true, 00:33:37.788 "claim_type": "exclusive_write", 00:33:37.788 "zoned": false, 00:33:37.788 "supported_io_types": { 00:33:37.788 "read": true, 00:33:37.788 "write": true, 00:33:37.788 "unmap": true, 00:33:37.788 "flush": true, 00:33:37.788 "reset": true, 00:33:37.788 "nvme_admin": false, 00:33:37.788 "nvme_io": false, 00:33:37.788 "nvme_io_md": false, 00:33:37.788 "write_zeroes": true, 00:33:37.788 "zcopy": true, 00:33:37.788 "get_zone_info": false, 00:33:37.788 "zone_management": false, 00:33:37.788 "zone_append": false, 00:33:37.788 "compare": false, 00:33:37.788 "compare_and_write": false, 00:33:37.788 "abort": true, 00:33:37.788 "seek_hole": false, 00:33:37.788 "seek_data": false, 00:33:37.788 "copy": true, 00:33:37.788 "nvme_iov_md": false 00:33:37.788 }, 00:33:37.788 "memory_domains": [ 00:33:37.788 { 00:33:37.788 "dma_device_id": "system", 00:33:37.788 "dma_device_type": 1 00:33:37.788 }, 00:33:37.788 { 00:33:37.788 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:37.788 "dma_device_type": 2 00:33:37.788 } 00:33:37.788 ], 00:33:37.788 "driver_specific": {} 00:33:37.788 } 00:33:37.788 ] 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:33:37.788 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:37.789 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:37.789 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:37.789 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:37.789 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:37.789 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:37.789 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:37.789 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:37.789 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:37.789 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:37.789 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:37.789 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:37.789 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:37.789 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:37.789 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:37.789 "name": "Existed_Raid", 00:33:37.789 "uuid": "a3bf0e8f-46bc-4dc5-812f-282cb289ab92", 00:33:37.789 "strip_size_kb": 64, 00:33:37.789 "state": "configuring", 00:33:37.789 "raid_level": "raid5f", 00:33:37.789 "superblock": true, 00:33:37.789 "num_base_bdevs": 4, 00:33:37.789 "num_base_bdevs_discovered": 1, 00:33:37.789 "num_base_bdevs_operational": 4, 00:33:37.789 "base_bdevs_list": [ 00:33:37.789 { 00:33:37.789 "name": "BaseBdev1", 00:33:37.789 "uuid": "0e46dddb-e2f4-4706-9dd8-026901d0168f", 00:33:37.789 "is_configured": true, 00:33:37.789 "data_offset": 2048, 00:33:37.789 "data_size": 63488 00:33:37.789 }, 00:33:37.789 { 00:33:37.789 "name": "BaseBdev2", 00:33:37.789 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:37.789 "is_configured": false, 00:33:37.789 "data_offset": 0, 00:33:37.789 "data_size": 0 00:33:37.789 }, 00:33:37.789 { 00:33:37.789 "name": "BaseBdev3", 00:33:37.789 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:37.789 "is_configured": false, 00:33:37.789 "data_offset": 0, 00:33:37.789 "data_size": 0 00:33:37.789 }, 00:33:37.789 { 00:33:37.789 "name": "BaseBdev4", 00:33:37.789 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:37.789 "is_configured": false, 00:33:37.789 "data_offset": 0, 00:33:37.789 "data_size": 0 00:33:37.789 } 00:33:37.789 ] 00:33:37.789 }' 00:33:37.789 13:53:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:37.789 13:53:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:38.047 [2024-11-20 13:53:45.037322] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:33:38.047 [2024-11-20 13:53:45.037379] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:38.047 [2024-11-20 13:53:45.045372] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:33:38.047 [2024-11-20 13:53:45.047315] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:33:38.047 [2024-11-20 13:53:45.047357] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:33:38.047 [2024-11-20 13:53:45.047366] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:33:38.047 [2024-11-20 13:53:45.047378] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:33:38.047 [2024-11-20 13:53:45.047385] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:33:38.047 [2024-11-20 13:53:45.047394] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:38.047 "name": "Existed_Raid", 00:33:38.047 "uuid": "1262db20-aeec-476c-bfc8-67081cb77f11", 00:33:38.047 "strip_size_kb": 64, 00:33:38.047 "state": "configuring", 00:33:38.047 "raid_level": "raid5f", 00:33:38.047 "superblock": true, 00:33:38.047 "num_base_bdevs": 4, 00:33:38.047 "num_base_bdevs_discovered": 1, 00:33:38.047 "num_base_bdevs_operational": 4, 00:33:38.047 "base_bdevs_list": [ 00:33:38.047 { 00:33:38.047 "name": "BaseBdev1", 00:33:38.047 "uuid": "0e46dddb-e2f4-4706-9dd8-026901d0168f", 00:33:38.047 "is_configured": true, 00:33:38.047 "data_offset": 2048, 00:33:38.047 "data_size": 63488 00:33:38.047 }, 00:33:38.047 { 00:33:38.047 "name": "BaseBdev2", 00:33:38.047 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:38.047 "is_configured": false, 00:33:38.047 "data_offset": 0, 00:33:38.047 "data_size": 0 00:33:38.047 }, 00:33:38.047 { 00:33:38.047 "name": "BaseBdev3", 00:33:38.047 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:38.047 "is_configured": false, 00:33:38.047 "data_offset": 0, 00:33:38.047 "data_size": 0 00:33:38.047 }, 00:33:38.047 { 00:33:38.047 "name": "BaseBdev4", 00:33:38.047 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:38.047 "is_configured": false, 00:33:38.047 "data_offset": 0, 00:33:38.047 "data_size": 0 00:33:38.047 } 00:33:38.047 ] 00:33:38.047 }' 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:38.047 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:38.613 [2024-11-20 13:53:45.414242] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:33:38.613 BaseBdev2 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:38.613 [ 00:33:38.613 { 00:33:38.613 "name": "BaseBdev2", 00:33:38.613 "aliases": [ 00:33:38.613 "96685066-b0e8-4202-9a69-aa896f656c06" 00:33:38.613 ], 00:33:38.613 "product_name": "Malloc disk", 00:33:38.613 "block_size": 512, 00:33:38.613 "num_blocks": 65536, 00:33:38.613 "uuid": "96685066-b0e8-4202-9a69-aa896f656c06", 00:33:38.613 "assigned_rate_limits": { 00:33:38.613 "rw_ios_per_sec": 0, 00:33:38.613 "rw_mbytes_per_sec": 0, 00:33:38.613 "r_mbytes_per_sec": 0, 00:33:38.613 "w_mbytes_per_sec": 0 00:33:38.613 }, 00:33:38.613 "claimed": true, 00:33:38.613 "claim_type": "exclusive_write", 00:33:38.613 "zoned": false, 00:33:38.613 "supported_io_types": { 00:33:38.613 "read": true, 00:33:38.613 "write": true, 00:33:38.613 "unmap": true, 00:33:38.613 "flush": true, 00:33:38.613 "reset": true, 00:33:38.613 "nvme_admin": false, 00:33:38.613 "nvme_io": false, 00:33:38.613 "nvme_io_md": false, 00:33:38.613 "write_zeroes": true, 00:33:38.613 "zcopy": true, 00:33:38.613 "get_zone_info": false, 00:33:38.613 "zone_management": false, 00:33:38.613 "zone_append": false, 00:33:38.613 "compare": false, 00:33:38.613 "compare_and_write": false, 00:33:38.613 "abort": true, 00:33:38.613 "seek_hole": false, 00:33:38.613 "seek_data": false, 00:33:38.613 "copy": true, 00:33:38.613 "nvme_iov_md": false 00:33:38.613 }, 00:33:38.613 "memory_domains": [ 00:33:38.613 { 00:33:38.613 "dma_device_id": "system", 00:33:38.613 "dma_device_type": 1 00:33:38.613 }, 00:33:38.613 { 00:33:38.613 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:38.613 "dma_device_type": 2 00:33:38.613 } 00:33:38.613 ], 00:33:38.613 "driver_specific": {} 00:33:38.613 } 00:33:38.613 ] 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:38.613 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:38.614 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:38.614 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:38.614 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:38.614 "name": "Existed_Raid", 00:33:38.614 "uuid": "1262db20-aeec-476c-bfc8-67081cb77f11", 00:33:38.614 "strip_size_kb": 64, 00:33:38.614 "state": "configuring", 00:33:38.614 "raid_level": "raid5f", 00:33:38.614 "superblock": true, 00:33:38.614 "num_base_bdevs": 4, 00:33:38.614 "num_base_bdevs_discovered": 2, 00:33:38.614 "num_base_bdevs_operational": 4, 00:33:38.614 "base_bdevs_list": [ 00:33:38.614 { 00:33:38.614 "name": "BaseBdev1", 00:33:38.614 "uuid": "0e46dddb-e2f4-4706-9dd8-026901d0168f", 00:33:38.614 "is_configured": true, 00:33:38.614 "data_offset": 2048, 00:33:38.614 "data_size": 63488 00:33:38.614 }, 00:33:38.614 { 00:33:38.614 "name": "BaseBdev2", 00:33:38.614 "uuid": "96685066-b0e8-4202-9a69-aa896f656c06", 00:33:38.614 "is_configured": true, 00:33:38.614 "data_offset": 2048, 00:33:38.614 "data_size": 63488 00:33:38.614 }, 00:33:38.614 { 00:33:38.614 "name": "BaseBdev3", 00:33:38.614 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:38.614 "is_configured": false, 00:33:38.614 "data_offset": 0, 00:33:38.614 "data_size": 0 00:33:38.614 }, 00:33:38.614 { 00:33:38.614 "name": "BaseBdev4", 00:33:38.614 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:38.614 "is_configured": false, 00:33:38.614 "data_offset": 0, 00:33:38.614 "data_size": 0 00:33:38.614 } 00:33:38.614 ] 00:33:38.614 }' 00:33:38.614 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:38.614 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:38.872 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:33:38.872 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:38.872 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:38.872 [2024-11-20 13:53:45.815811] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:33:38.872 BaseBdev3 00:33:38.872 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:38.872 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:33:38.872 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:33:38.872 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:33:38.872 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:33:38.872 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:33:38.872 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:33:38.872 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:33:38.872 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:38.872 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:38.872 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:38.872 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:33:38.872 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:38.872 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:38.872 [ 00:33:38.872 { 00:33:38.872 "name": "BaseBdev3", 00:33:38.872 "aliases": [ 00:33:38.872 "5fc23a3a-3548-4f56-ae60-4cd800536070" 00:33:38.872 ], 00:33:38.872 "product_name": "Malloc disk", 00:33:38.872 "block_size": 512, 00:33:38.872 "num_blocks": 65536, 00:33:38.872 "uuid": "5fc23a3a-3548-4f56-ae60-4cd800536070", 00:33:38.872 "assigned_rate_limits": { 00:33:38.872 "rw_ios_per_sec": 0, 00:33:38.872 "rw_mbytes_per_sec": 0, 00:33:38.872 "r_mbytes_per_sec": 0, 00:33:38.872 "w_mbytes_per_sec": 0 00:33:38.872 }, 00:33:38.872 "claimed": true, 00:33:38.872 "claim_type": "exclusive_write", 00:33:38.872 "zoned": false, 00:33:38.872 "supported_io_types": { 00:33:38.872 "read": true, 00:33:38.872 "write": true, 00:33:38.872 "unmap": true, 00:33:38.872 "flush": true, 00:33:38.872 "reset": true, 00:33:38.872 "nvme_admin": false, 00:33:38.872 "nvme_io": false, 00:33:38.872 "nvme_io_md": false, 00:33:38.872 "write_zeroes": true, 00:33:38.872 "zcopy": true, 00:33:38.872 "get_zone_info": false, 00:33:38.872 "zone_management": false, 00:33:38.872 "zone_append": false, 00:33:38.872 "compare": false, 00:33:38.872 "compare_and_write": false, 00:33:38.873 "abort": true, 00:33:38.873 "seek_hole": false, 00:33:38.873 "seek_data": false, 00:33:38.873 "copy": true, 00:33:38.873 "nvme_iov_md": false 00:33:38.873 }, 00:33:38.873 "memory_domains": [ 00:33:38.873 { 00:33:38.873 "dma_device_id": "system", 00:33:38.873 "dma_device_type": 1 00:33:38.873 }, 00:33:38.873 { 00:33:38.873 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:38.873 "dma_device_type": 2 00:33:38.873 } 00:33:38.873 ], 00:33:38.873 "driver_specific": {} 00:33:38.873 } 00:33:38.873 ] 00:33:38.873 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:38.873 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:33:38.873 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:33:38.873 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:33:38.873 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:33:38.873 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:38.873 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:38.873 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:38.873 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:38.873 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:38.873 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:38.873 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:38.873 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:38.873 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:38.873 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:38.873 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:38.873 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:38.873 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:38.873 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:38.873 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:38.873 "name": "Existed_Raid", 00:33:38.873 "uuid": "1262db20-aeec-476c-bfc8-67081cb77f11", 00:33:38.873 "strip_size_kb": 64, 00:33:38.873 "state": "configuring", 00:33:38.873 "raid_level": "raid5f", 00:33:38.873 "superblock": true, 00:33:38.873 "num_base_bdevs": 4, 00:33:38.873 "num_base_bdevs_discovered": 3, 00:33:38.873 "num_base_bdevs_operational": 4, 00:33:38.873 "base_bdevs_list": [ 00:33:38.873 { 00:33:38.873 "name": "BaseBdev1", 00:33:38.873 "uuid": "0e46dddb-e2f4-4706-9dd8-026901d0168f", 00:33:38.873 "is_configured": true, 00:33:38.873 "data_offset": 2048, 00:33:38.873 "data_size": 63488 00:33:38.873 }, 00:33:38.873 { 00:33:38.873 "name": "BaseBdev2", 00:33:38.873 "uuid": "96685066-b0e8-4202-9a69-aa896f656c06", 00:33:38.873 "is_configured": true, 00:33:38.873 "data_offset": 2048, 00:33:38.873 "data_size": 63488 00:33:38.873 }, 00:33:38.873 { 00:33:38.873 "name": "BaseBdev3", 00:33:38.873 "uuid": "5fc23a3a-3548-4f56-ae60-4cd800536070", 00:33:38.873 "is_configured": true, 00:33:38.873 "data_offset": 2048, 00:33:38.873 "data_size": 63488 00:33:38.873 }, 00:33:38.873 { 00:33:38.873 "name": "BaseBdev4", 00:33:38.873 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:38.873 "is_configured": false, 00:33:38.873 "data_offset": 0, 00:33:38.873 "data_size": 0 00:33:38.873 } 00:33:38.873 ] 00:33:38.873 }' 00:33:38.873 13:53:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:38.873 13:53:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:39.131 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:33:39.131 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:39.131 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:39.389 [2024-11-20 13:53:46.201553] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:33:39.389 [2024-11-20 13:53:46.202031] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:33:39.389 [2024-11-20 13:53:46.202129] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:33:39.389 BaseBdev4 00:33:39.389 [2024-11-20 13:53:46.202441] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:33:39.389 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:39.389 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:33:39.389 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:33:39.389 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:33:39.389 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:33:39.389 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:33:39.389 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:33:39.389 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:33:39.389 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:39.389 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:39.389 [2024-11-20 13:53:46.207457] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:33:39.389 [2024-11-20 13:53:46.207479] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:33:39.389 [2024-11-20 13:53:46.207723] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:39.389 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:39.389 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:33:39.389 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:39.389 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:39.389 [ 00:33:39.389 { 00:33:39.389 "name": "BaseBdev4", 00:33:39.389 "aliases": [ 00:33:39.389 "d0f8dd23-c2b9-4544-be8f-29420bda6bec" 00:33:39.389 ], 00:33:39.389 "product_name": "Malloc disk", 00:33:39.389 "block_size": 512, 00:33:39.389 "num_blocks": 65536, 00:33:39.389 "uuid": "d0f8dd23-c2b9-4544-be8f-29420bda6bec", 00:33:39.389 "assigned_rate_limits": { 00:33:39.389 "rw_ios_per_sec": 0, 00:33:39.389 "rw_mbytes_per_sec": 0, 00:33:39.389 "r_mbytes_per_sec": 0, 00:33:39.389 "w_mbytes_per_sec": 0 00:33:39.389 }, 00:33:39.389 "claimed": true, 00:33:39.389 "claim_type": "exclusive_write", 00:33:39.389 "zoned": false, 00:33:39.389 "supported_io_types": { 00:33:39.389 "read": true, 00:33:39.389 "write": true, 00:33:39.389 "unmap": true, 00:33:39.389 "flush": true, 00:33:39.389 "reset": true, 00:33:39.389 "nvme_admin": false, 00:33:39.389 "nvme_io": false, 00:33:39.389 "nvme_io_md": false, 00:33:39.389 "write_zeroes": true, 00:33:39.389 "zcopy": true, 00:33:39.389 "get_zone_info": false, 00:33:39.389 "zone_management": false, 00:33:39.389 "zone_append": false, 00:33:39.389 "compare": false, 00:33:39.389 "compare_and_write": false, 00:33:39.390 "abort": true, 00:33:39.390 "seek_hole": false, 00:33:39.390 "seek_data": false, 00:33:39.390 "copy": true, 00:33:39.390 "nvme_iov_md": false 00:33:39.390 }, 00:33:39.390 "memory_domains": [ 00:33:39.390 { 00:33:39.390 "dma_device_id": "system", 00:33:39.390 "dma_device_type": 1 00:33:39.390 }, 00:33:39.390 { 00:33:39.390 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:39.390 "dma_device_type": 2 00:33:39.390 } 00:33:39.390 ], 00:33:39.390 "driver_specific": {} 00:33:39.390 } 00:33:39.390 ] 00:33:39.390 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:39.390 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:33:39.390 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:33:39.390 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:33:39.390 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:33:39.390 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:39.390 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:39.390 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:39.390 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:39.390 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:39.390 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:39.390 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:39.390 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:39.390 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:39.390 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:39.390 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:39.390 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:39.390 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:39.390 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:39.390 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:39.390 "name": "Existed_Raid", 00:33:39.390 "uuid": "1262db20-aeec-476c-bfc8-67081cb77f11", 00:33:39.390 "strip_size_kb": 64, 00:33:39.390 "state": "online", 00:33:39.390 "raid_level": "raid5f", 00:33:39.390 "superblock": true, 00:33:39.390 "num_base_bdevs": 4, 00:33:39.390 "num_base_bdevs_discovered": 4, 00:33:39.390 "num_base_bdevs_operational": 4, 00:33:39.390 "base_bdevs_list": [ 00:33:39.390 { 00:33:39.390 "name": "BaseBdev1", 00:33:39.390 "uuid": "0e46dddb-e2f4-4706-9dd8-026901d0168f", 00:33:39.390 "is_configured": true, 00:33:39.390 "data_offset": 2048, 00:33:39.390 "data_size": 63488 00:33:39.390 }, 00:33:39.390 { 00:33:39.390 "name": "BaseBdev2", 00:33:39.390 "uuid": "96685066-b0e8-4202-9a69-aa896f656c06", 00:33:39.390 "is_configured": true, 00:33:39.390 "data_offset": 2048, 00:33:39.390 "data_size": 63488 00:33:39.390 }, 00:33:39.390 { 00:33:39.390 "name": "BaseBdev3", 00:33:39.390 "uuid": "5fc23a3a-3548-4f56-ae60-4cd800536070", 00:33:39.390 "is_configured": true, 00:33:39.390 "data_offset": 2048, 00:33:39.390 "data_size": 63488 00:33:39.390 }, 00:33:39.390 { 00:33:39.390 "name": "BaseBdev4", 00:33:39.390 "uuid": "d0f8dd23-c2b9-4544-be8f-29420bda6bec", 00:33:39.390 "is_configured": true, 00:33:39.390 "data_offset": 2048, 00:33:39.390 "data_size": 63488 00:33:39.390 } 00:33:39.390 ] 00:33:39.390 }' 00:33:39.390 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:39.390 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:39.648 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:33:39.648 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:33:39.648 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:33:39.648 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:33:39.648 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:33:39.648 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:33:39.648 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:33:39.648 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:33:39.648 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:39.648 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:39.648 [2024-11-20 13:53:46.561768] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:33:39.648 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:39.648 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:33:39.648 "name": "Existed_Raid", 00:33:39.648 "aliases": [ 00:33:39.648 "1262db20-aeec-476c-bfc8-67081cb77f11" 00:33:39.648 ], 00:33:39.648 "product_name": "Raid Volume", 00:33:39.648 "block_size": 512, 00:33:39.648 "num_blocks": 190464, 00:33:39.648 "uuid": "1262db20-aeec-476c-bfc8-67081cb77f11", 00:33:39.648 "assigned_rate_limits": { 00:33:39.648 "rw_ios_per_sec": 0, 00:33:39.648 "rw_mbytes_per_sec": 0, 00:33:39.648 "r_mbytes_per_sec": 0, 00:33:39.648 "w_mbytes_per_sec": 0 00:33:39.648 }, 00:33:39.648 "claimed": false, 00:33:39.648 "zoned": false, 00:33:39.648 "supported_io_types": { 00:33:39.648 "read": true, 00:33:39.648 "write": true, 00:33:39.648 "unmap": false, 00:33:39.648 "flush": false, 00:33:39.648 "reset": true, 00:33:39.648 "nvme_admin": false, 00:33:39.648 "nvme_io": false, 00:33:39.648 "nvme_io_md": false, 00:33:39.648 "write_zeroes": true, 00:33:39.648 "zcopy": false, 00:33:39.648 "get_zone_info": false, 00:33:39.648 "zone_management": false, 00:33:39.648 "zone_append": false, 00:33:39.648 "compare": false, 00:33:39.648 "compare_and_write": false, 00:33:39.648 "abort": false, 00:33:39.649 "seek_hole": false, 00:33:39.649 "seek_data": false, 00:33:39.649 "copy": false, 00:33:39.649 "nvme_iov_md": false 00:33:39.649 }, 00:33:39.649 "driver_specific": { 00:33:39.649 "raid": { 00:33:39.649 "uuid": "1262db20-aeec-476c-bfc8-67081cb77f11", 00:33:39.649 "strip_size_kb": 64, 00:33:39.649 "state": "online", 00:33:39.649 "raid_level": "raid5f", 00:33:39.649 "superblock": true, 00:33:39.649 "num_base_bdevs": 4, 00:33:39.649 "num_base_bdevs_discovered": 4, 00:33:39.649 "num_base_bdevs_operational": 4, 00:33:39.649 "base_bdevs_list": [ 00:33:39.649 { 00:33:39.649 "name": "BaseBdev1", 00:33:39.649 "uuid": "0e46dddb-e2f4-4706-9dd8-026901d0168f", 00:33:39.649 "is_configured": true, 00:33:39.649 "data_offset": 2048, 00:33:39.649 "data_size": 63488 00:33:39.649 }, 00:33:39.649 { 00:33:39.649 "name": "BaseBdev2", 00:33:39.649 "uuid": "96685066-b0e8-4202-9a69-aa896f656c06", 00:33:39.649 "is_configured": true, 00:33:39.649 "data_offset": 2048, 00:33:39.649 "data_size": 63488 00:33:39.649 }, 00:33:39.649 { 00:33:39.649 "name": "BaseBdev3", 00:33:39.649 "uuid": "5fc23a3a-3548-4f56-ae60-4cd800536070", 00:33:39.649 "is_configured": true, 00:33:39.649 "data_offset": 2048, 00:33:39.649 "data_size": 63488 00:33:39.649 }, 00:33:39.649 { 00:33:39.649 "name": "BaseBdev4", 00:33:39.649 "uuid": "d0f8dd23-c2b9-4544-be8f-29420bda6bec", 00:33:39.649 "is_configured": true, 00:33:39.649 "data_offset": 2048, 00:33:39.649 "data_size": 63488 00:33:39.649 } 00:33:39.649 ] 00:33:39.649 } 00:33:39.649 } 00:33:39.649 }' 00:33:39.649 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:33:39.649 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:33:39.649 BaseBdev2 00:33:39.649 BaseBdev3 00:33:39.649 BaseBdev4' 00:33:39.649 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:39.649 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:33:39.649 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:39.649 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:39.649 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:33:39.649 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:39.649 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:39.649 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:39.649 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:33:39.649 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:33:39.649 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:39.649 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:33:39.649 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:39.649 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:39.649 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:39.649 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:39.907 [2024-11-20 13:53:46.781647] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:39.907 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:39.907 "name": "Existed_Raid", 00:33:39.907 "uuid": "1262db20-aeec-476c-bfc8-67081cb77f11", 00:33:39.907 "strip_size_kb": 64, 00:33:39.907 "state": "online", 00:33:39.907 "raid_level": "raid5f", 00:33:39.907 "superblock": true, 00:33:39.907 "num_base_bdevs": 4, 00:33:39.907 "num_base_bdevs_discovered": 3, 00:33:39.907 "num_base_bdevs_operational": 3, 00:33:39.907 "base_bdevs_list": [ 00:33:39.908 { 00:33:39.908 "name": null, 00:33:39.908 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:39.908 "is_configured": false, 00:33:39.908 "data_offset": 0, 00:33:39.908 "data_size": 63488 00:33:39.908 }, 00:33:39.908 { 00:33:39.908 "name": "BaseBdev2", 00:33:39.908 "uuid": "96685066-b0e8-4202-9a69-aa896f656c06", 00:33:39.908 "is_configured": true, 00:33:39.908 "data_offset": 2048, 00:33:39.908 "data_size": 63488 00:33:39.908 }, 00:33:39.908 { 00:33:39.908 "name": "BaseBdev3", 00:33:39.908 "uuid": "5fc23a3a-3548-4f56-ae60-4cd800536070", 00:33:39.908 "is_configured": true, 00:33:39.908 "data_offset": 2048, 00:33:39.908 "data_size": 63488 00:33:39.908 }, 00:33:39.908 { 00:33:39.908 "name": "BaseBdev4", 00:33:39.908 "uuid": "d0f8dd23-c2b9-4544-be8f-29420bda6bec", 00:33:39.908 "is_configured": true, 00:33:39.908 "data_offset": 2048, 00:33:39.908 "data_size": 63488 00:33:39.908 } 00:33:39.908 ] 00:33:39.908 }' 00:33:39.908 13:53:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:39.908 13:53:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:40.166 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:33:40.166 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:33:40.166 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:40.166 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:40.166 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:40.166 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:33:40.166 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:40.166 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:33:40.166 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:33:40.166 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:33:40.166 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:40.166 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:40.166 [2024-11-20 13:53:47.197600] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:33:40.166 [2024-11-20 13:53:47.197904] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:33:40.424 [2024-11-20 13:53:47.260762] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:40.424 [2024-11-20 13:53:47.300803] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:40.424 [2024-11-20 13:53:47.403686] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:33:40.424 [2024-11-20 13:53:47.403827] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:40.424 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:40.683 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:40.683 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:33:40.683 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:33:40.683 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:33:40.683 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:33:40.683 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:33:40.683 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:33:40.683 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:40.683 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:40.683 BaseBdev2 00:33:40.683 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:40.683 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:33:40.683 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:33:40.683 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:33:40.683 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:33:40.683 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:33:40.683 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:33:40.683 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:33:40.683 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:40.683 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:40.683 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:40.683 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:33:40.683 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:40.683 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:40.683 [ 00:33:40.683 { 00:33:40.683 "name": "BaseBdev2", 00:33:40.683 "aliases": [ 00:33:40.683 "0524708c-4eef-4dac-8830-22f3a27f869a" 00:33:40.683 ], 00:33:40.683 "product_name": "Malloc disk", 00:33:40.683 "block_size": 512, 00:33:40.683 "num_blocks": 65536, 00:33:40.683 "uuid": "0524708c-4eef-4dac-8830-22f3a27f869a", 00:33:40.683 "assigned_rate_limits": { 00:33:40.683 "rw_ios_per_sec": 0, 00:33:40.683 "rw_mbytes_per_sec": 0, 00:33:40.683 "r_mbytes_per_sec": 0, 00:33:40.683 "w_mbytes_per_sec": 0 00:33:40.683 }, 00:33:40.683 "claimed": false, 00:33:40.683 "zoned": false, 00:33:40.683 "supported_io_types": { 00:33:40.683 "read": true, 00:33:40.683 "write": true, 00:33:40.683 "unmap": true, 00:33:40.683 "flush": true, 00:33:40.683 "reset": true, 00:33:40.683 "nvme_admin": false, 00:33:40.683 "nvme_io": false, 00:33:40.683 "nvme_io_md": false, 00:33:40.683 "write_zeroes": true, 00:33:40.683 "zcopy": true, 00:33:40.683 "get_zone_info": false, 00:33:40.683 "zone_management": false, 00:33:40.683 "zone_append": false, 00:33:40.683 "compare": false, 00:33:40.683 "compare_and_write": false, 00:33:40.683 "abort": true, 00:33:40.683 "seek_hole": false, 00:33:40.683 "seek_data": false, 00:33:40.683 "copy": true, 00:33:40.683 "nvme_iov_md": false 00:33:40.683 }, 00:33:40.683 "memory_domains": [ 00:33:40.683 { 00:33:40.683 "dma_device_id": "system", 00:33:40.683 "dma_device_type": 1 00:33:40.683 }, 00:33:40.683 { 00:33:40.683 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:40.684 "dma_device_type": 2 00:33:40.684 } 00:33:40.684 ], 00:33:40.684 "driver_specific": {} 00:33:40.684 } 00:33:40.684 ] 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:40.684 BaseBdev3 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:40.684 [ 00:33:40.684 { 00:33:40.684 "name": "BaseBdev3", 00:33:40.684 "aliases": [ 00:33:40.684 "9d41bbf5-4e8a-4160-b895-b37037d83a8e" 00:33:40.684 ], 00:33:40.684 "product_name": "Malloc disk", 00:33:40.684 "block_size": 512, 00:33:40.684 "num_blocks": 65536, 00:33:40.684 "uuid": "9d41bbf5-4e8a-4160-b895-b37037d83a8e", 00:33:40.684 "assigned_rate_limits": { 00:33:40.684 "rw_ios_per_sec": 0, 00:33:40.684 "rw_mbytes_per_sec": 0, 00:33:40.684 "r_mbytes_per_sec": 0, 00:33:40.684 "w_mbytes_per_sec": 0 00:33:40.684 }, 00:33:40.684 "claimed": false, 00:33:40.684 "zoned": false, 00:33:40.684 "supported_io_types": { 00:33:40.684 "read": true, 00:33:40.684 "write": true, 00:33:40.684 "unmap": true, 00:33:40.684 "flush": true, 00:33:40.684 "reset": true, 00:33:40.684 "nvme_admin": false, 00:33:40.684 "nvme_io": false, 00:33:40.684 "nvme_io_md": false, 00:33:40.684 "write_zeroes": true, 00:33:40.684 "zcopy": true, 00:33:40.684 "get_zone_info": false, 00:33:40.684 "zone_management": false, 00:33:40.684 "zone_append": false, 00:33:40.684 "compare": false, 00:33:40.684 "compare_and_write": false, 00:33:40.684 "abort": true, 00:33:40.684 "seek_hole": false, 00:33:40.684 "seek_data": false, 00:33:40.684 "copy": true, 00:33:40.684 "nvme_iov_md": false 00:33:40.684 }, 00:33:40.684 "memory_domains": [ 00:33:40.684 { 00:33:40.684 "dma_device_id": "system", 00:33:40.684 "dma_device_type": 1 00:33:40.684 }, 00:33:40.684 { 00:33:40.684 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:40.684 "dma_device_type": 2 00:33:40.684 } 00:33:40.684 ], 00:33:40.684 "driver_specific": {} 00:33:40.684 } 00:33:40.684 ] 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:40.684 BaseBdev4 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:40.684 [ 00:33:40.684 { 00:33:40.684 "name": "BaseBdev4", 00:33:40.684 "aliases": [ 00:33:40.684 "f14c0289-f045-4eca-be7b-8fbf1a7e0747" 00:33:40.684 ], 00:33:40.684 "product_name": "Malloc disk", 00:33:40.684 "block_size": 512, 00:33:40.684 "num_blocks": 65536, 00:33:40.684 "uuid": "f14c0289-f045-4eca-be7b-8fbf1a7e0747", 00:33:40.684 "assigned_rate_limits": { 00:33:40.684 "rw_ios_per_sec": 0, 00:33:40.684 "rw_mbytes_per_sec": 0, 00:33:40.684 "r_mbytes_per_sec": 0, 00:33:40.684 "w_mbytes_per_sec": 0 00:33:40.684 }, 00:33:40.684 "claimed": false, 00:33:40.684 "zoned": false, 00:33:40.684 "supported_io_types": { 00:33:40.684 "read": true, 00:33:40.684 "write": true, 00:33:40.684 "unmap": true, 00:33:40.684 "flush": true, 00:33:40.684 "reset": true, 00:33:40.684 "nvme_admin": false, 00:33:40.684 "nvme_io": false, 00:33:40.684 "nvme_io_md": false, 00:33:40.684 "write_zeroes": true, 00:33:40.684 "zcopy": true, 00:33:40.684 "get_zone_info": false, 00:33:40.684 "zone_management": false, 00:33:40.684 "zone_append": false, 00:33:40.684 "compare": false, 00:33:40.684 "compare_and_write": false, 00:33:40.684 "abort": true, 00:33:40.684 "seek_hole": false, 00:33:40.684 "seek_data": false, 00:33:40.684 "copy": true, 00:33:40.684 "nvme_iov_md": false 00:33:40.684 }, 00:33:40.684 "memory_domains": [ 00:33:40.684 { 00:33:40.684 "dma_device_id": "system", 00:33:40.684 "dma_device_type": 1 00:33:40.684 }, 00:33:40.684 { 00:33:40.684 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:40.684 "dma_device_type": 2 00:33:40.684 } 00:33:40.684 ], 00:33:40.684 "driver_specific": {} 00:33:40.684 } 00:33:40.684 ] 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:40.684 [2024-11-20 13:53:47.694664] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:33:40.684 [2024-11-20 13:53:47.694812] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:33:40.684 [2024-11-20 13:53:47.694956] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:33:40.684 [2024-11-20 13:53:47.697032] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:33:40.684 [2024-11-20 13:53:47.697160] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:40.684 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:40.685 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:40.685 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:40.685 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:40.685 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:40.685 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:40.685 "name": "Existed_Raid", 00:33:40.685 "uuid": "fd49def1-87a2-4378-89fb-7ba2b459896f", 00:33:40.685 "strip_size_kb": 64, 00:33:40.685 "state": "configuring", 00:33:40.685 "raid_level": "raid5f", 00:33:40.685 "superblock": true, 00:33:40.685 "num_base_bdevs": 4, 00:33:40.685 "num_base_bdevs_discovered": 3, 00:33:40.685 "num_base_bdevs_operational": 4, 00:33:40.685 "base_bdevs_list": [ 00:33:40.685 { 00:33:40.685 "name": "BaseBdev1", 00:33:40.685 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:40.685 "is_configured": false, 00:33:40.685 "data_offset": 0, 00:33:40.685 "data_size": 0 00:33:40.685 }, 00:33:40.685 { 00:33:40.685 "name": "BaseBdev2", 00:33:40.685 "uuid": "0524708c-4eef-4dac-8830-22f3a27f869a", 00:33:40.685 "is_configured": true, 00:33:40.685 "data_offset": 2048, 00:33:40.685 "data_size": 63488 00:33:40.685 }, 00:33:40.685 { 00:33:40.685 "name": "BaseBdev3", 00:33:40.685 "uuid": "9d41bbf5-4e8a-4160-b895-b37037d83a8e", 00:33:40.685 "is_configured": true, 00:33:40.685 "data_offset": 2048, 00:33:40.685 "data_size": 63488 00:33:40.685 }, 00:33:40.685 { 00:33:40.685 "name": "BaseBdev4", 00:33:40.685 "uuid": "f14c0289-f045-4eca-be7b-8fbf1a7e0747", 00:33:40.685 "is_configured": true, 00:33:40.685 "data_offset": 2048, 00:33:40.685 "data_size": 63488 00:33:40.685 } 00:33:40.685 ] 00:33:40.685 }' 00:33:40.685 13:53:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:40.685 13:53:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:41.251 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:33:41.251 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:41.251 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:41.251 [2024-11-20 13:53:48.054767] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:33:41.251 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:41.251 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:33:41.251 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:41.251 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:41.251 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:41.251 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:41.251 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:41.251 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:41.251 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:41.251 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:41.251 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:41.251 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:41.251 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:41.251 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:41.251 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:41.251 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:41.251 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:41.251 "name": "Existed_Raid", 00:33:41.251 "uuid": "fd49def1-87a2-4378-89fb-7ba2b459896f", 00:33:41.251 "strip_size_kb": 64, 00:33:41.251 "state": "configuring", 00:33:41.251 "raid_level": "raid5f", 00:33:41.251 "superblock": true, 00:33:41.251 "num_base_bdevs": 4, 00:33:41.251 "num_base_bdevs_discovered": 2, 00:33:41.251 "num_base_bdevs_operational": 4, 00:33:41.251 "base_bdevs_list": [ 00:33:41.251 { 00:33:41.251 "name": "BaseBdev1", 00:33:41.251 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:41.251 "is_configured": false, 00:33:41.251 "data_offset": 0, 00:33:41.251 "data_size": 0 00:33:41.251 }, 00:33:41.251 { 00:33:41.251 "name": null, 00:33:41.251 "uuid": "0524708c-4eef-4dac-8830-22f3a27f869a", 00:33:41.251 "is_configured": false, 00:33:41.251 "data_offset": 0, 00:33:41.251 "data_size": 63488 00:33:41.251 }, 00:33:41.251 { 00:33:41.251 "name": "BaseBdev3", 00:33:41.251 "uuid": "9d41bbf5-4e8a-4160-b895-b37037d83a8e", 00:33:41.251 "is_configured": true, 00:33:41.251 "data_offset": 2048, 00:33:41.251 "data_size": 63488 00:33:41.251 }, 00:33:41.251 { 00:33:41.251 "name": "BaseBdev4", 00:33:41.251 "uuid": "f14c0289-f045-4eca-be7b-8fbf1a7e0747", 00:33:41.251 "is_configured": true, 00:33:41.251 "data_offset": 2048, 00:33:41.251 "data_size": 63488 00:33:41.251 } 00:33:41.251 ] 00:33:41.251 }' 00:33:41.251 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:41.251 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:41.509 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:41.509 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:33:41.509 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:41.509 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:41.509 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:41.509 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:33:41.509 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:33:41.509 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:41.509 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:41.509 [2024-11-20 13:53:48.427460] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:33:41.509 BaseBdev1 00:33:41.509 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:41.509 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:33:41.509 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:33:41.509 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:33:41.509 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:33:41.509 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:33:41.509 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:33:41.509 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:33:41.509 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:41.509 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:41.509 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:41.509 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:33:41.509 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:41.510 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:41.510 [ 00:33:41.510 { 00:33:41.510 "name": "BaseBdev1", 00:33:41.510 "aliases": [ 00:33:41.510 "7c837dff-3dd9-4d21-a77c-72f8f1e58dd1" 00:33:41.510 ], 00:33:41.510 "product_name": "Malloc disk", 00:33:41.510 "block_size": 512, 00:33:41.510 "num_blocks": 65536, 00:33:41.510 "uuid": "7c837dff-3dd9-4d21-a77c-72f8f1e58dd1", 00:33:41.510 "assigned_rate_limits": { 00:33:41.510 "rw_ios_per_sec": 0, 00:33:41.510 "rw_mbytes_per_sec": 0, 00:33:41.510 "r_mbytes_per_sec": 0, 00:33:41.510 "w_mbytes_per_sec": 0 00:33:41.510 }, 00:33:41.510 "claimed": true, 00:33:41.510 "claim_type": "exclusive_write", 00:33:41.510 "zoned": false, 00:33:41.510 "supported_io_types": { 00:33:41.510 "read": true, 00:33:41.510 "write": true, 00:33:41.510 "unmap": true, 00:33:41.510 "flush": true, 00:33:41.510 "reset": true, 00:33:41.510 "nvme_admin": false, 00:33:41.510 "nvme_io": false, 00:33:41.510 "nvme_io_md": false, 00:33:41.510 "write_zeroes": true, 00:33:41.510 "zcopy": true, 00:33:41.510 "get_zone_info": false, 00:33:41.510 "zone_management": false, 00:33:41.510 "zone_append": false, 00:33:41.510 "compare": false, 00:33:41.510 "compare_and_write": false, 00:33:41.510 "abort": true, 00:33:41.510 "seek_hole": false, 00:33:41.510 "seek_data": false, 00:33:41.510 "copy": true, 00:33:41.510 "nvme_iov_md": false 00:33:41.510 }, 00:33:41.510 "memory_domains": [ 00:33:41.510 { 00:33:41.510 "dma_device_id": "system", 00:33:41.510 "dma_device_type": 1 00:33:41.510 }, 00:33:41.510 { 00:33:41.510 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:41.510 "dma_device_type": 2 00:33:41.510 } 00:33:41.510 ], 00:33:41.510 "driver_specific": {} 00:33:41.510 } 00:33:41.510 ] 00:33:41.510 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:41.510 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:33:41.510 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:33:41.510 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:41.510 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:41.510 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:41.510 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:41.510 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:41.510 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:41.510 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:41.510 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:41.510 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:41.510 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:41.510 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:41.510 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:41.510 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:41.510 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:41.510 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:41.510 "name": "Existed_Raid", 00:33:41.510 "uuid": "fd49def1-87a2-4378-89fb-7ba2b459896f", 00:33:41.510 "strip_size_kb": 64, 00:33:41.510 "state": "configuring", 00:33:41.510 "raid_level": "raid5f", 00:33:41.510 "superblock": true, 00:33:41.510 "num_base_bdevs": 4, 00:33:41.510 "num_base_bdevs_discovered": 3, 00:33:41.510 "num_base_bdevs_operational": 4, 00:33:41.510 "base_bdevs_list": [ 00:33:41.510 { 00:33:41.510 "name": "BaseBdev1", 00:33:41.510 "uuid": "7c837dff-3dd9-4d21-a77c-72f8f1e58dd1", 00:33:41.510 "is_configured": true, 00:33:41.510 "data_offset": 2048, 00:33:41.510 "data_size": 63488 00:33:41.510 }, 00:33:41.510 { 00:33:41.510 "name": null, 00:33:41.510 "uuid": "0524708c-4eef-4dac-8830-22f3a27f869a", 00:33:41.510 "is_configured": false, 00:33:41.510 "data_offset": 0, 00:33:41.510 "data_size": 63488 00:33:41.510 }, 00:33:41.510 { 00:33:41.510 "name": "BaseBdev3", 00:33:41.510 "uuid": "9d41bbf5-4e8a-4160-b895-b37037d83a8e", 00:33:41.510 "is_configured": true, 00:33:41.510 "data_offset": 2048, 00:33:41.510 "data_size": 63488 00:33:41.510 }, 00:33:41.510 { 00:33:41.510 "name": "BaseBdev4", 00:33:41.510 "uuid": "f14c0289-f045-4eca-be7b-8fbf1a7e0747", 00:33:41.510 "is_configured": true, 00:33:41.510 "data_offset": 2048, 00:33:41.510 "data_size": 63488 00:33:41.510 } 00:33:41.510 ] 00:33:41.510 }' 00:33:41.510 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:41.510 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:41.769 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:41.769 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:41.769 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:41.769 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:33:41.769 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:41.769 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:33:41.769 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:33:41.769 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:41.769 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:41.769 [2024-11-20 13:53:48.811655] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:33:41.769 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:41.769 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:33:41.769 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:41.769 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:41.769 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:41.769 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:41.769 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:41.769 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:41.769 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:41.769 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:41.769 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:41.769 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:41.769 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:41.769 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:41.769 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:42.027 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:42.027 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:42.027 "name": "Existed_Raid", 00:33:42.027 "uuid": "fd49def1-87a2-4378-89fb-7ba2b459896f", 00:33:42.027 "strip_size_kb": 64, 00:33:42.027 "state": "configuring", 00:33:42.027 "raid_level": "raid5f", 00:33:42.027 "superblock": true, 00:33:42.027 "num_base_bdevs": 4, 00:33:42.027 "num_base_bdevs_discovered": 2, 00:33:42.027 "num_base_bdevs_operational": 4, 00:33:42.027 "base_bdevs_list": [ 00:33:42.027 { 00:33:42.027 "name": "BaseBdev1", 00:33:42.027 "uuid": "7c837dff-3dd9-4d21-a77c-72f8f1e58dd1", 00:33:42.027 "is_configured": true, 00:33:42.027 "data_offset": 2048, 00:33:42.027 "data_size": 63488 00:33:42.027 }, 00:33:42.027 { 00:33:42.027 "name": null, 00:33:42.027 "uuid": "0524708c-4eef-4dac-8830-22f3a27f869a", 00:33:42.027 "is_configured": false, 00:33:42.027 "data_offset": 0, 00:33:42.027 "data_size": 63488 00:33:42.027 }, 00:33:42.027 { 00:33:42.027 "name": null, 00:33:42.027 "uuid": "9d41bbf5-4e8a-4160-b895-b37037d83a8e", 00:33:42.027 "is_configured": false, 00:33:42.027 "data_offset": 0, 00:33:42.027 "data_size": 63488 00:33:42.027 }, 00:33:42.027 { 00:33:42.027 "name": "BaseBdev4", 00:33:42.027 "uuid": "f14c0289-f045-4eca-be7b-8fbf1a7e0747", 00:33:42.027 "is_configured": true, 00:33:42.027 "data_offset": 2048, 00:33:42.027 "data_size": 63488 00:33:42.027 } 00:33:42.027 ] 00:33:42.027 }' 00:33:42.027 13:53:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:42.027 13:53:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:42.290 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:42.290 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:42.290 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:42.290 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:33:42.290 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:42.290 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:33:42.290 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:33:42.290 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:42.290 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:42.290 [2024-11-20 13:53:49.151703] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:33:42.290 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:42.290 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:33:42.290 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:42.291 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:42.291 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:42.291 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:42.291 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:42.291 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:42.291 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:42.291 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:42.291 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:42.291 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:42.291 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:42.291 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:42.291 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:42.291 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:42.291 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:42.291 "name": "Existed_Raid", 00:33:42.291 "uuid": "fd49def1-87a2-4378-89fb-7ba2b459896f", 00:33:42.291 "strip_size_kb": 64, 00:33:42.291 "state": "configuring", 00:33:42.291 "raid_level": "raid5f", 00:33:42.291 "superblock": true, 00:33:42.291 "num_base_bdevs": 4, 00:33:42.291 "num_base_bdevs_discovered": 3, 00:33:42.291 "num_base_bdevs_operational": 4, 00:33:42.291 "base_bdevs_list": [ 00:33:42.292 { 00:33:42.292 "name": "BaseBdev1", 00:33:42.292 "uuid": "7c837dff-3dd9-4d21-a77c-72f8f1e58dd1", 00:33:42.292 "is_configured": true, 00:33:42.292 "data_offset": 2048, 00:33:42.292 "data_size": 63488 00:33:42.292 }, 00:33:42.292 { 00:33:42.292 "name": null, 00:33:42.292 "uuid": "0524708c-4eef-4dac-8830-22f3a27f869a", 00:33:42.292 "is_configured": false, 00:33:42.292 "data_offset": 0, 00:33:42.292 "data_size": 63488 00:33:42.292 }, 00:33:42.292 { 00:33:42.292 "name": "BaseBdev3", 00:33:42.292 "uuid": "9d41bbf5-4e8a-4160-b895-b37037d83a8e", 00:33:42.292 "is_configured": true, 00:33:42.292 "data_offset": 2048, 00:33:42.292 "data_size": 63488 00:33:42.292 }, 00:33:42.292 { 00:33:42.292 "name": "BaseBdev4", 00:33:42.292 "uuid": "f14c0289-f045-4eca-be7b-8fbf1a7e0747", 00:33:42.292 "is_configured": true, 00:33:42.292 "data_offset": 2048, 00:33:42.292 "data_size": 63488 00:33:42.292 } 00:33:42.292 ] 00:33:42.292 }' 00:33:42.292 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:42.292 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:42.554 [2024-11-20 13:53:49.511809] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:42.554 "name": "Existed_Raid", 00:33:42.554 "uuid": "fd49def1-87a2-4378-89fb-7ba2b459896f", 00:33:42.554 "strip_size_kb": 64, 00:33:42.554 "state": "configuring", 00:33:42.554 "raid_level": "raid5f", 00:33:42.554 "superblock": true, 00:33:42.554 "num_base_bdevs": 4, 00:33:42.554 "num_base_bdevs_discovered": 2, 00:33:42.554 "num_base_bdevs_operational": 4, 00:33:42.554 "base_bdevs_list": [ 00:33:42.554 { 00:33:42.554 "name": null, 00:33:42.554 "uuid": "7c837dff-3dd9-4d21-a77c-72f8f1e58dd1", 00:33:42.554 "is_configured": false, 00:33:42.554 "data_offset": 0, 00:33:42.554 "data_size": 63488 00:33:42.554 }, 00:33:42.554 { 00:33:42.554 "name": null, 00:33:42.554 "uuid": "0524708c-4eef-4dac-8830-22f3a27f869a", 00:33:42.554 "is_configured": false, 00:33:42.554 "data_offset": 0, 00:33:42.554 "data_size": 63488 00:33:42.554 }, 00:33:42.554 { 00:33:42.554 "name": "BaseBdev3", 00:33:42.554 "uuid": "9d41bbf5-4e8a-4160-b895-b37037d83a8e", 00:33:42.554 "is_configured": true, 00:33:42.554 "data_offset": 2048, 00:33:42.554 "data_size": 63488 00:33:42.554 }, 00:33:42.554 { 00:33:42.554 "name": "BaseBdev4", 00:33:42.554 "uuid": "f14c0289-f045-4eca-be7b-8fbf1a7e0747", 00:33:42.554 "is_configured": true, 00:33:42.554 "data_offset": 2048, 00:33:42.554 "data_size": 63488 00:33:42.554 } 00:33:42.554 ] 00:33:42.554 }' 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:42.554 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:43.119 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:43.119 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:33:43.119 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:43.119 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:43.119 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:43.119 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:33:43.119 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:33:43.119 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:43.119 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:43.119 [2024-11-20 13:53:49.941764] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:33:43.119 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:43.119 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:33:43.119 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:43.119 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:43.119 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:43.119 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:43.119 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:43.119 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:43.119 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:43.120 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:43.120 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:43.120 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:43.120 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:43.120 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:43.120 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:43.120 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:43.120 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:43.120 "name": "Existed_Raid", 00:33:43.120 "uuid": "fd49def1-87a2-4378-89fb-7ba2b459896f", 00:33:43.120 "strip_size_kb": 64, 00:33:43.120 "state": "configuring", 00:33:43.120 "raid_level": "raid5f", 00:33:43.120 "superblock": true, 00:33:43.120 "num_base_bdevs": 4, 00:33:43.120 "num_base_bdevs_discovered": 3, 00:33:43.120 "num_base_bdevs_operational": 4, 00:33:43.120 "base_bdevs_list": [ 00:33:43.120 { 00:33:43.120 "name": null, 00:33:43.120 "uuid": "7c837dff-3dd9-4d21-a77c-72f8f1e58dd1", 00:33:43.120 "is_configured": false, 00:33:43.120 "data_offset": 0, 00:33:43.120 "data_size": 63488 00:33:43.120 }, 00:33:43.120 { 00:33:43.120 "name": "BaseBdev2", 00:33:43.120 "uuid": "0524708c-4eef-4dac-8830-22f3a27f869a", 00:33:43.120 "is_configured": true, 00:33:43.120 "data_offset": 2048, 00:33:43.120 "data_size": 63488 00:33:43.120 }, 00:33:43.120 { 00:33:43.120 "name": "BaseBdev3", 00:33:43.120 "uuid": "9d41bbf5-4e8a-4160-b895-b37037d83a8e", 00:33:43.120 "is_configured": true, 00:33:43.120 "data_offset": 2048, 00:33:43.120 "data_size": 63488 00:33:43.120 }, 00:33:43.120 { 00:33:43.120 "name": "BaseBdev4", 00:33:43.120 "uuid": "f14c0289-f045-4eca-be7b-8fbf1a7e0747", 00:33:43.120 "is_configured": true, 00:33:43.120 "data_offset": 2048, 00:33:43.120 "data_size": 63488 00:33:43.120 } 00:33:43.120 ] 00:33:43.120 }' 00:33:43.120 13:53:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:43.120 13:53:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 7c837dff-3dd9-4d21-a77c-72f8f1e58dd1 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:43.378 [2024-11-20 13:53:50.370189] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:33:43.378 [2024-11-20 13:53:50.370384] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:33:43.378 [2024-11-20 13:53:50.370396] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:33:43.378 NewBaseBdev 00:33:43.378 [2024-11-20 13:53:50.370612] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:43.378 [2024-11-20 13:53:50.374538] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:33:43.378 [2024-11-20 13:53:50.374558] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:33:43.378 [2024-11-20 13:53:50.374684] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:43.378 [ 00:33:43.378 { 00:33:43.378 "name": "NewBaseBdev", 00:33:43.378 "aliases": [ 00:33:43.378 "7c837dff-3dd9-4d21-a77c-72f8f1e58dd1" 00:33:43.378 ], 00:33:43.378 "product_name": "Malloc disk", 00:33:43.378 "block_size": 512, 00:33:43.378 "num_blocks": 65536, 00:33:43.378 "uuid": "7c837dff-3dd9-4d21-a77c-72f8f1e58dd1", 00:33:43.378 "assigned_rate_limits": { 00:33:43.378 "rw_ios_per_sec": 0, 00:33:43.378 "rw_mbytes_per_sec": 0, 00:33:43.378 "r_mbytes_per_sec": 0, 00:33:43.378 "w_mbytes_per_sec": 0 00:33:43.378 }, 00:33:43.378 "claimed": true, 00:33:43.378 "claim_type": "exclusive_write", 00:33:43.378 "zoned": false, 00:33:43.378 "supported_io_types": { 00:33:43.378 "read": true, 00:33:43.378 "write": true, 00:33:43.378 "unmap": true, 00:33:43.378 "flush": true, 00:33:43.378 "reset": true, 00:33:43.378 "nvme_admin": false, 00:33:43.378 "nvme_io": false, 00:33:43.378 "nvme_io_md": false, 00:33:43.378 "write_zeroes": true, 00:33:43.378 "zcopy": true, 00:33:43.378 "get_zone_info": false, 00:33:43.378 "zone_management": false, 00:33:43.378 "zone_append": false, 00:33:43.378 "compare": false, 00:33:43.378 "compare_and_write": false, 00:33:43.378 "abort": true, 00:33:43.378 "seek_hole": false, 00:33:43.378 "seek_data": false, 00:33:43.378 "copy": true, 00:33:43.378 "nvme_iov_md": false 00:33:43.378 }, 00:33:43.378 "memory_domains": [ 00:33:43.378 { 00:33:43.378 "dma_device_id": "system", 00:33:43.378 "dma_device_type": 1 00:33:43.378 }, 00:33:43.378 { 00:33:43.378 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:43.378 "dma_device_type": 2 00:33:43.378 } 00:33:43.378 ], 00:33:43.378 "driver_specific": {} 00:33:43.378 } 00:33:43.378 ] 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:43.378 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:43.636 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:43.636 "name": "Existed_Raid", 00:33:43.636 "uuid": "fd49def1-87a2-4378-89fb-7ba2b459896f", 00:33:43.636 "strip_size_kb": 64, 00:33:43.636 "state": "online", 00:33:43.636 "raid_level": "raid5f", 00:33:43.636 "superblock": true, 00:33:43.636 "num_base_bdevs": 4, 00:33:43.636 "num_base_bdevs_discovered": 4, 00:33:43.636 "num_base_bdevs_operational": 4, 00:33:43.636 "base_bdevs_list": [ 00:33:43.636 { 00:33:43.636 "name": "NewBaseBdev", 00:33:43.636 "uuid": "7c837dff-3dd9-4d21-a77c-72f8f1e58dd1", 00:33:43.636 "is_configured": true, 00:33:43.636 "data_offset": 2048, 00:33:43.636 "data_size": 63488 00:33:43.636 }, 00:33:43.636 { 00:33:43.636 "name": "BaseBdev2", 00:33:43.636 "uuid": "0524708c-4eef-4dac-8830-22f3a27f869a", 00:33:43.636 "is_configured": true, 00:33:43.636 "data_offset": 2048, 00:33:43.636 "data_size": 63488 00:33:43.636 }, 00:33:43.636 { 00:33:43.636 "name": "BaseBdev3", 00:33:43.636 "uuid": "9d41bbf5-4e8a-4160-b895-b37037d83a8e", 00:33:43.636 "is_configured": true, 00:33:43.636 "data_offset": 2048, 00:33:43.636 "data_size": 63488 00:33:43.636 }, 00:33:43.636 { 00:33:43.636 "name": "BaseBdev4", 00:33:43.636 "uuid": "f14c0289-f045-4eca-be7b-8fbf1a7e0747", 00:33:43.636 "is_configured": true, 00:33:43.636 "data_offset": 2048, 00:33:43.636 "data_size": 63488 00:33:43.636 } 00:33:43.637 ] 00:33:43.637 }' 00:33:43.637 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:43.637 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:43.896 [2024-11-20 13:53:50.735542] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:33:43.896 "name": "Existed_Raid", 00:33:43.896 "aliases": [ 00:33:43.896 "fd49def1-87a2-4378-89fb-7ba2b459896f" 00:33:43.896 ], 00:33:43.896 "product_name": "Raid Volume", 00:33:43.896 "block_size": 512, 00:33:43.896 "num_blocks": 190464, 00:33:43.896 "uuid": "fd49def1-87a2-4378-89fb-7ba2b459896f", 00:33:43.896 "assigned_rate_limits": { 00:33:43.896 "rw_ios_per_sec": 0, 00:33:43.896 "rw_mbytes_per_sec": 0, 00:33:43.896 "r_mbytes_per_sec": 0, 00:33:43.896 "w_mbytes_per_sec": 0 00:33:43.896 }, 00:33:43.896 "claimed": false, 00:33:43.896 "zoned": false, 00:33:43.896 "supported_io_types": { 00:33:43.896 "read": true, 00:33:43.896 "write": true, 00:33:43.896 "unmap": false, 00:33:43.896 "flush": false, 00:33:43.896 "reset": true, 00:33:43.896 "nvme_admin": false, 00:33:43.896 "nvme_io": false, 00:33:43.896 "nvme_io_md": false, 00:33:43.896 "write_zeroes": true, 00:33:43.896 "zcopy": false, 00:33:43.896 "get_zone_info": false, 00:33:43.896 "zone_management": false, 00:33:43.896 "zone_append": false, 00:33:43.896 "compare": false, 00:33:43.896 "compare_and_write": false, 00:33:43.896 "abort": false, 00:33:43.896 "seek_hole": false, 00:33:43.896 "seek_data": false, 00:33:43.896 "copy": false, 00:33:43.896 "nvme_iov_md": false 00:33:43.896 }, 00:33:43.896 "driver_specific": { 00:33:43.896 "raid": { 00:33:43.896 "uuid": "fd49def1-87a2-4378-89fb-7ba2b459896f", 00:33:43.896 "strip_size_kb": 64, 00:33:43.896 "state": "online", 00:33:43.896 "raid_level": "raid5f", 00:33:43.896 "superblock": true, 00:33:43.896 "num_base_bdevs": 4, 00:33:43.896 "num_base_bdevs_discovered": 4, 00:33:43.896 "num_base_bdevs_operational": 4, 00:33:43.896 "base_bdevs_list": [ 00:33:43.896 { 00:33:43.896 "name": "NewBaseBdev", 00:33:43.896 "uuid": "7c837dff-3dd9-4d21-a77c-72f8f1e58dd1", 00:33:43.896 "is_configured": true, 00:33:43.896 "data_offset": 2048, 00:33:43.896 "data_size": 63488 00:33:43.896 }, 00:33:43.896 { 00:33:43.896 "name": "BaseBdev2", 00:33:43.896 "uuid": "0524708c-4eef-4dac-8830-22f3a27f869a", 00:33:43.896 "is_configured": true, 00:33:43.896 "data_offset": 2048, 00:33:43.896 "data_size": 63488 00:33:43.896 }, 00:33:43.896 { 00:33:43.896 "name": "BaseBdev3", 00:33:43.896 "uuid": "9d41bbf5-4e8a-4160-b895-b37037d83a8e", 00:33:43.896 "is_configured": true, 00:33:43.896 "data_offset": 2048, 00:33:43.896 "data_size": 63488 00:33:43.896 }, 00:33:43.896 { 00:33:43.896 "name": "BaseBdev4", 00:33:43.896 "uuid": "f14c0289-f045-4eca-be7b-8fbf1a7e0747", 00:33:43.896 "is_configured": true, 00:33:43.896 "data_offset": 2048, 00:33:43.896 "data_size": 63488 00:33:43.896 } 00:33:43.896 ] 00:33:43.896 } 00:33:43.896 } 00:33:43.896 }' 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:33:43.896 BaseBdev2 00:33:43.896 BaseBdev3 00:33:43.896 BaseBdev4' 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:43.896 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:43.897 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:43.897 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:44.155 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:33:44.155 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:33:44.155 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:33:44.155 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:44.155 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:44.155 [2024-11-20 13:53:50.963363] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:33:44.155 [2024-11-20 13:53:50.963391] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:33:44.155 [2024-11-20 13:53:50.963467] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:33:44.155 [2024-11-20 13:53:50.963728] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:33:44.155 [2024-11-20 13:53:50.963737] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:33:44.155 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:44.155 13:53:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 81018 00:33:44.155 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 81018 ']' 00:33:44.155 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 81018 00:33:44.155 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:33:44.155 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:33:44.155 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 81018 00:33:44.155 killing process with pid 81018 00:33:44.155 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:33:44.155 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:33:44.155 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 81018' 00:33:44.155 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 81018 00:33:44.155 [2024-11-20 13:53:50.994952] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:33:44.155 13:53:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 81018 00:33:44.155 [2024-11-20 13:53:51.205669] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:33:45.177 ************************************ 00:33:45.177 END TEST raid5f_state_function_test_sb 00:33:45.177 ************************************ 00:33:45.177 13:53:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:33:45.177 00:33:45.177 real 0m8.426s 00:33:45.177 user 0m13.390s 00:33:45.177 sys 0m1.505s 00:33:45.177 13:53:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:33:45.177 13:53:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:33:45.177 13:53:51 bdev_raid -- bdev/bdev_raid.sh@988 -- # run_test raid5f_superblock_test raid_superblock_test raid5f 4 00:33:45.177 13:53:51 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:33:45.177 13:53:51 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:33:45.177 13:53:51 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:33:45.177 ************************************ 00:33:45.177 START TEST raid5f_superblock_test 00:33:45.177 ************************************ 00:33:45.177 13:53:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid5f 4 00:33:45.177 13:53:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid5f 00:33:45.177 13:53:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:33:45.177 13:53:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:33:45.177 13:53:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:33:45.177 13:53:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:33:45.177 13:53:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:33:45.177 13:53:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:33:45.177 13:53:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:33:45.177 13:53:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:33:45.177 13:53:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:33:45.177 13:53:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:33:45.177 13:53:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:33:45.177 13:53:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:33:45.177 13:53:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid5f '!=' raid1 ']' 00:33:45.177 13:53:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:33:45.177 13:53:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:33:45.177 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:33:45.177 13:53:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=81656 00:33:45.177 13:53:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 81656 00:33:45.177 13:53:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 81656 ']' 00:33:45.177 13:53:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:33:45.177 13:53:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:33:45.177 13:53:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:33:45.177 13:53:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:33:45.177 13:53:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:33:45.177 13:53:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:45.177 [2024-11-20 13:53:51.944669] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:33:45.177 [2024-11-20 13:53:51.944950] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid81656 ] 00:33:45.177 [2024-11-20 13:53:52.099722] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:33:45.177 [2024-11-20 13:53:52.203057] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:33:45.436 [2024-11-20 13:53:52.328121] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:33:45.436 [2024-11-20 13:53:52.328304] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.003 malloc1 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.003 [2024-11-20 13:53:52.836431] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:33:46.003 [2024-11-20 13:53:52.836603] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:46.003 [2024-11-20 13:53:52.836642] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:33:46.003 [2024-11-20 13:53:52.836694] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:46.003 [2024-11-20 13:53:52.838536] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:46.003 [2024-11-20 13:53:52.838646] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:33:46.003 pt1 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.003 malloc2 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.003 [2024-11-20 13:53:52.874363] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:33:46.003 [2024-11-20 13:53:52.874410] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:46.003 [2024-11-20 13:53:52.874431] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:33:46.003 [2024-11-20 13:53:52.874439] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:46.003 [2024-11-20 13:53:52.876545] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:46.003 [2024-11-20 13:53:52.876599] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:33:46.003 pt2 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.003 malloc3 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.003 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.003 [2024-11-20 13:53:52.926252] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:33:46.003 [2024-11-20 13:53:52.926300] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:46.003 [2024-11-20 13:53:52.926320] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:33:46.003 [2024-11-20 13:53:52.926327] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:46.003 [2024-11-20 13:53:52.928177] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:46.003 [2024-11-20 13:53:52.928207] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:33:46.004 pt3 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.004 malloc4 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.004 [2024-11-20 13:53:52.960157] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:33:46.004 [2024-11-20 13:53:52.960205] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:46.004 [2024-11-20 13:53:52.960220] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:33:46.004 [2024-11-20 13:53:52.960227] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:46.004 [2024-11-20 13:53:52.962087] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:46.004 [2024-11-20 13:53:52.962114] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:33:46.004 pt4 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.004 [2024-11-20 13:53:52.968194] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:33:46.004 [2024-11-20 13:53:52.969851] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:33:46.004 [2024-11-20 13:53:52.970039] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:33:46.004 [2024-11-20 13:53:52.970086] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:33:46.004 [2024-11-20 13:53:52.970250] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:33:46.004 [2024-11-20 13:53:52.970263] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:33:46.004 [2024-11-20 13:53:52.970490] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:33:46.004 [2024-11-20 13:53:52.974803] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:33:46.004 [2024-11-20 13:53:52.974823] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:33:46.004 [2024-11-20 13:53:52.975006] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:46.004 13:53:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:46.004 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:46.004 "name": "raid_bdev1", 00:33:46.004 "uuid": "c63ab43f-9a61-4278-82cc-962656385e73", 00:33:46.004 "strip_size_kb": 64, 00:33:46.004 "state": "online", 00:33:46.004 "raid_level": "raid5f", 00:33:46.004 "superblock": true, 00:33:46.004 "num_base_bdevs": 4, 00:33:46.004 "num_base_bdevs_discovered": 4, 00:33:46.004 "num_base_bdevs_operational": 4, 00:33:46.004 "base_bdevs_list": [ 00:33:46.004 { 00:33:46.004 "name": "pt1", 00:33:46.004 "uuid": "00000000-0000-0000-0000-000000000001", 00:33:46.004 "is_configured": true, 00:33:46.004 "data_offset": 2048, 00:33:46.004 "data_size": 63488 00:33:46.004 }, 00:33:46.004 { 00:33:46.004 "name": "pt2", 00:33:46.004 "uuid": "00000000-0000-0000-0000-000000000002", 00:33:46.004 "is_configured": true, 00:33:46.004 "data_offset": 2048, 00:33:46.004 "data_size": 63488 00:33:46.004 }, 00:33:46.004 { 00:33:46.004 "name": "pt3", 00:33:46.004 "uuid": "00000000-0000-0000-0000-000000000003", 00:33:46.004 "is_configured": true, 00:33:46.004 "data_offset": 2048, 00:33:46.004 "data_size": 63488 00:33:46.004 }, 00:33:46.004 { 00:33:46.004 "name": "pt4", 00:33:46.004 "uuid": "00000000-0000-0000-0000-000000000004", 00:33:46.004 "is_configured": true, 00:33:46.004 "data_offset": 2048, 00:33:46.004 "data_size": 63488 00:33:46.004 } 00:33:46.004 ] 00:33:46.004 }' 00:33:46.004 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:46.004 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.261 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:33:46.261 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:33:46.261 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:33:46.261 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:33:46.261 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:33:46.261 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:33:46.261 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:33:46.261 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:33:46.261 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.261 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.261 [2024-11-20 13:53:53.283976] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:33:46.261 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:46.261 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:33:46.261 "name": "raid_bdev1", 00:33:46.261 "aliases": [ 00:33:46.261 "c63ab43f-9a61-4278-82cc-962656385e73" 00:33:46.261 ], 00:33:46.261 "product_name": "Raid Volume", 00:33:46.261 "block_size": 512, 00:33:46.261 "num_blocks": 190464, 00:33:46.261 "uuid": "c63ab43f-9a61-4278-82cc-962656385e73", 00:33:46.261 "assigned_rate_limits": { 00:33:46.261 "rw_ios_per_sec": 0, 00:33:46.261 "rw_mbytes_per_sec": 0, 00:33:46.261 "r_mbytes_per_sec": 0, 00:33:46.261 "w_mbytes_per_sec": 0 00:33:46.261 }, 00:33:46.261 "claimed": false, 00:33:46.261 "zoned": false, 00:33:46.261 "supported_io_types": { 00:33:46.261 "read": true, 00:33:46.261 "write": true, 00:33:46.261 "unmap": false, 00:33:46.261 "flush": false, 00:33:46.261 "reset": true, 00:33:46.261 "nvme_admin": false, 00:33:46.261 "nvme_io": false, 00:33:46.261 "nvme_io_md": false, 00:33:46.261 "write_zeroes": true, 00:33:46.261 "zcopy": false, 00:33:46.261 "get_zone_info": false, 00:33:46.261 "zone_management": false, 00:33:46.261 "zone_append": false, 00:33:46.261 "compare": false, 00:33:46.261 "compare_and_write": false, 00:33:46.261 "abort": false, 00:33:46.261 "seek_hole": false, 00:33:46.261 "seek_data": false, 00:33:46.261 "copy": false, 00:33:46.261 "nvme_iov_md": false 00:33:46.261 }, 00:33:46.261 "driver_specific": { 00:33:46.261 "raid": { 00:33:46.261 "uuid": "c63ab43f-9a61-4278-82cc-962656385e73", 00:33:46.261 "strip_size_kb": 64, 00:33:46.261 "state": "online", 00:33:46.261 "raid_level": "raid5f", 00:33:46.261 "superblock": true, 00:33:46.261 "num_base_bdevs": 4, 00:33:46.261 "num_base_bdevs_discovered": 4, 00:33:46.261 "num_base_bdevs_operational": 4, 00:33:46.261 "base_bdevs_list": [ 00:33:46.261 { 00:33:46.261 "name": "pt1", 00:33:46.261 "uuid": "00000000-0000-0000-0000-000000000001", 00:33:46.261 "is_configured": true, 00:33:46.261 "data_offset": 2048, 00:33:46.261 "data_size": 63488 00:33:46.261 }, 00:33:46.261 { 00:33:46.261 "name": "pt2", 00:33:46.261 "uuid": "00000000-0000-0000-0000-000000000002", 00:33:46.261 "is_configured": true, 00:33:46.261 "data_offset": 2048, 00:33:46.261 "data_size": 63488 00:33:46.261 }, 00:33:46.261 { 00:33:46.261 "name": "pt3", 00:33:46.261 "uuid": "00000000-0000-0000-0000-000000000003", 00:33:46.261 "is_configured": true, 00:33:46.261 "data_offset": 2048, 00:33:46.261 "data_size": 63488 00:33:46.261 }, 00:33:46.261 { 00:33:46.261 "name": "pt4", 00:33:46.261 "uuid": "00000000-0000-0000-0000-000000000004", 00:33:46.261 "is_configured": true, 00:33:46.261 "data_offset": 2048, 00:33:46.261 "data_size": 63488 00:33:46.261 } 00:33:46.261 ] 00:33:46.261 } 00:33:46.261 } 00:33:46.261 }' 00:33:46.261 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:33:46.518 pt2 00:33:46.518 pt3 00:33:46.518 pt4' 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:33:46.518 [2024-11-20 13:53:53.515947] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=c63ab43f-9a61-4278-82cc-962656385e73 00:33:46.518 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z c63ab43f-9a61-4278-82cc-962656385e73 ']' 00:33:46.519 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:33:46.519 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.519 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.519 [2024-11-20 13:53:53.543797] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:33:46.519 [2024-11-20 13:53:53.543905] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:33:46.519 [2024-11-20 13:53:53.544006] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:33:46.519 [2024-11-20 13:53:53.544092] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:33:46.519 [2024-11-20 13:53:53.544105] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:33:46.519 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:46.519 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:46.519 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.519 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.519 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:33:46.519 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.777 [2024-11-20 13:53:53.659878] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:33:46.777 [2024-11-20 13:53:53.661625] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:33:46.777 [2024-11-20 13:53:53.661671] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:33:46.777 [2024-11-20 13:53:53.661700] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:33:46.777 [2024-11-20 13:53:53.661747] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:33:46.777 [2024-11-20 13:53:53.661797] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:33:46.777 [2024-11-20 13:53:53.661813] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:33:46.777 [2024-11-20 13:53:53.661828] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:33:46.777 [2024-11-20 13:53:53.661838] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:33:46.777 [2024-11-20 13:53:53.661849] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:33:46.777 request: 00:33:46.777 { 00:33:46.777 "name": "raid_bdev1", 00:33:46.777 "raid_level": "raid5f", 00:33:46.777 "base_bdevs": [ 00:33:46.777 "malloc1", 00:33:46.777 "malloc2", 00:33:46.777 "malloc3", 00:33:46.777 "malloc4" 00:33:46.777 ], 00:33:46.777 "strip_size_kb": 64, 00:33:46.777 "superblock": false, 00:33:46.777 "method": "bdev_raid_create", 00:33:46.777 "req_id": 1 00:33:46.777 } 00:33:46.777 Got JSON-RPC error response 00:33:46.777 response: 00:33:46.777 { 00:33:46.777 "code": -17, 00:33:46.777 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:33:46.777 } 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.777 [2024-11-20 13:53:53.695853] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:33:46.777 [2024-11-20 13:53:53.695908] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:46.777 [2024-11-20 13:53:53.695926] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:33:46.777 [2024-11-20 13:53:53.695935] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:46.777 [2024-11-20 13:53:53.697943] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:46.777 [2024-11-20 13:53:53.697989] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:33:46.777 [2024-11-20 13:53:53.698070] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:33:46.777 [2024-11-20 13:53:53.698120] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:33:46.777 pt1 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 4 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:46.777 "name": "raid_bdev1", 00:33:46.777 "uuid": "c63ab43f-9a61-4278-82cc-962656385e73", 00:33:46.777 "strip_size_kb": 64, 00:33:46.777 "state": "configuring", 00:33:46.777 "raid_level": "raid5f", 00:33:46.777 "superblock": true, 00:33:46.777 "num_base_bdevs": 4, 00:33:46.777 "num_base_bdevs_discovered": 1, 00:33:46.777 "num_base_bdevs_operational": 4, 00:33:46.777 "base_bdevs_list": [ 00:33:46.777 { 00:33:46.777 "name": "pt1", 00:33:46.777 "uuid": "00000000-0000-0000-0000-000000000001", 00:33:46.777 "is_configured": true, 00:33:46.777 "data_offset": 2048, 00:33:46.777 "data_size": 63488 00:33:46.777 }, 00:33:46.777 { 00:33:46.777 "name": null, 00:33:46.777 "uuid": "00000000-0000-0000-0000-000000000002", 00:33:46.777 "is_configured": false, 00:33:46.777 "data_offset": 2048, 00:33:46.777 "data_size": 63488 00:33:46.777 }, 00:33:46.777 { 00:33:46.777 "name": null, 00:33:46.777 "uuid": "00000000-0000-0000-0000-000000000003", 00:33:46.777 "is_configured": false, 00:33:46.777 "data_offset": 2048, 00:33:46.777 "data_size": 63488 00:33:46.777 }, 00:33:46.777 { 00:33:46.777 "name": null, 00:33:46.777 "uuid": "00000000-0000-0000-0000-000000000004", 00:33:46.777 "is_configured": false, 00:33:46.777 "data_offset": 2048, 00:33:46.777 "data_size": 63488 00:33:46.777 } 00:33:46.777 ] 00:33:46.777 }' 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:46.777 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:47.035 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:33:47.035 13:53:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:33:47.035 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:47.035 13:53:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:47.035 [2024-11-20 13:53:54.003922] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:33:47.035 [2024-11-20 13:53:54.004006] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:47.035 [2024-11-20 13:53:54.004024] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:33:47.035 [2024-11-20 13:53:54.004034] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:47.035 [2024-11-20 13:53:54.004432] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:47.035 [2024-11-20 13:53:54.004446] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:33:47.035 [2024-11-20 13:53:54.004516] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:33:47.035 [2024-11-20 13:53:54.004544] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:33:47.035 pt2 00:33:47.035 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:47.035 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:33:47.035 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:47.035 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:47.035 [2024-11-20 13:53:54.011917] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:33:47.035 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:47.035 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 4 00:33:47.035 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:47.035 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:47.035 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:47.035 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:47.035 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:47.035 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:47.035 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:47.035 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:47.035 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:47.035 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:47.035 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:47.035 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:47.035 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:47.035 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:47.035 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:47.035 "name": "raid_bdev1", 00:33:47.035 "uuid": "c63ab43f-9a61-4278-82cc-962656385e73", 00:33:47.035 "strip_size_kb": 64, 00:33:47.035 "state": "configuring", 00:33:47.035 "raid_level": "raid5f", 00:33:47.035 "superblock": true, 00:33:47.035 "num_base_bdevs": 4, 00:33:47.035 "num_base_bdevs_discovered": 1, 00:33:47.035 "num_base_bdevs_operational": 4, 00:33:47.035 "base_bdevs_list": [ 00:33:47.035 { 00:33:47.035 "name": "pt1", 00:33:47.035 "uuid": "00000000-0000-0000-0000-000000000001", 00:33:47.035 "is_configured": true, 00:33:47.035 "data_offset": 2048, 00:33:47.035 "data_size": 63488 00:33:47.035 }, 00:33:47.035 { 00:33:47.035 "name": null, 00:33:47.035 "uuid": "00000000-0000-0000-0000-000000000002", 00:33:47.035 "is_configured": false, 00:33:47.035 "data_offset": 0, 00:33:47.035 "data_size": 63488 00:33:47.035 }, 00:33:47.035 { 00:33:47.035 "name": null, 00:33:47.035 "uuid": "00000000-0000-0000-0000-000000000003", 00:33:47.035 "is_configured": false, 00:33:47.035 "data_offset": 2048, 00:33:47.035 "data_size": 63488 00:33:47.035 }, 00:33:47.035 { 00:33:47.035 "name": null, 00:33:47.035 "uuid": "00000000-0000-0000-0000-000000000004", 00:33:47.035 "is_configured": false, 00:33:47.035 "data_offset": 2048, 00:33:47.035 "data_size": 63488 00:33:47.035 } 00:33:47.035 ] 00:33:47.035 }' 00:33:47.035 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:47.035 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:47.293 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:33:47.293 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:33:47.293 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:33:47.293 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:47.293 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:47.293 [2024-11-20 13:53:54.319968] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:33:47.293 [2024-11-20 13:53:54.320146] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:47.293 [2024-11-20 13:53:54.320221] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:33:47.293 [2024-11-20 13:53:54.320267] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:47.293 [2024-11-20 13:53:54.320706] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:47.293 [2024-11-20 13:53:54.320797] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:33:47.293 [2024-11-20 13:53:54.320925] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:33:47.293 [2024-11-20 13:53:54.321001] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:33:47.293 pt2 00:33:47.293 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:47.293 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:33:47.293 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:33:47.293 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:33:47.293 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:47.293 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:47.293 [2024-11-20 13:53:54.327935] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:33:47.293 [2024-11-20 13:53:54.327974] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:47.293 [2024-11-20 13:53:54.328004] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:33:47.293 [2024-11-20 13:53:54.328011] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:47.293 [2024-11-20 13:53:54.328317] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:47.293 [2024-11-20 13:53:54.328338] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:33:47.293 [2024-11-20 13:53:54.328388] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:33:47.293 [2024-11-20 13:53:54.328405] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:33:47.293 pt3 00:33:47.293 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:47.293 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:33:47.293 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:33:47.293 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:33:47.294 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:47.294 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:47.294 [2024-11-20 13:53:54.335914] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:33:47.294 [2024-11-20 13:53:54.335947] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:47.294 [2024-11-20 13:53:54.335960] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:33:47.294 [2024-11-20 13:53:54.335966] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:47.294 [2024-11-20 13:53:54.336274] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:47.294 [2024-11-20 13:53:54.336294] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:33:47.294 [2024-11-20 13:53:54.336340] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:33:47.294 [2024-11-20 13:53:54.336355] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:33:47.294 [2024-11-20 13:53:54.336468] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:33:47.294 [2024-11-20 13:53:54.336479] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:33:47.294 [2024-11-20 13:53:54.336690] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:33:47.294 [2024-11-20 13:53:54.340457] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:33:47.294 [2024-11-20 13:53:54.340474] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:33:47.294 [2024-11-20 13:53:54.340616] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:47.294 pt4 00:33:47.294 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:47.294 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:33:47.294 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:33:47.294 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:33:47.294 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:47.294 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:47.294 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:47.294 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:47.294 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:47.294 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:47.294 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:47.294 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:47.294 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:47.294 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:47.294 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:47.294 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:47.294 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:47.551 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:47.551 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:47.551 "name": "raid_bdev1", 00:33:47.551 "uuid": "c63ab43f-9a61-4278-82cc-962656385e73", 00:33:47.551 "strip_size_kb": 64, 00:33:47.551 "state": "online", 00:33:47.551 "raid_level": "raid5f", 00:33:47.551 "superblock": true, 00:33:47.551 "num_base_bdevs": 4, 00:33:47.551 "num_base_bdevs_discovered": 4, 00:33:47.551 "num_base_bdevs_operational": 4, 00:33:47.551 "base_bdevs_list": [ 00:33:47.551 { 00:33:47.551 "name": "pt1", 00:33:47.551 "uuid": "00000000-0000-0000-0000-000000000001", 00:33:47.551 "is_configured": true, 00:33:47.551 "data_offset": 2048, 00:33:47.551 "data_size": 63488 00:33:47.551 }, 00:33:47.551 { 00:33:47.551 "name": "pt2", 00:33:47.551 "uuid": "00000000-0000-0000-0000-000000000002", 00:33:47.551 "is_configured": true, 00:33:47.551 "data_offset": 2048, 00:33:47.551 "data_size": 63488 00:33:47.551 }, 00:33:47.551 { 00:33:47.551 "name": "pt3", 00:33:47.551 "uuid": "00000000-0000-0000-0000-000000000003", 00:33:47.551 "is_configured": true, 00:33:47.551 "data_offset": 2048, 00:33:47.551 "data_size": 63488 00:33:47.551 }, 00:33:47.551 { 00:33:47.551 "name": "pt4", 00:33:47.551 "uuid": "00000000-0000-0000-0000-000000000004", 00:33:47.551 "is_configured": true, 00:33:47.551 "data_offset": 2048, 00:33:47.551 "data_size": 63488 00:33:47.551 } 00:33:47.551 ] 00:33:47.551 }' 00:33:47.551 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:47.551 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:47.809 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:33:47.809 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:33:47.809 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:33:47.809 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:33:47.809 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:33:47.809 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:33:47.809 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:33:47.809 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:47.809 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:47.809 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:33:47.809 [2024-11-20 13:53:54.645456] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:33:47.809 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:47.809 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:33:47.809 "name": "raid_bdev1", 00:33:47.809 "aliases": [ 00:33:47.809 "c63ab43f-9a61-4278-82cc-962656385e73" 00:33:47.809 ], 00:33:47.809 "product_name": "Raid Volume", 00:33:47.809 "block_size": 512, 00:33:47.809 "num_blocks": 190464, 00:33:47.809 "uuid": "c63ab43f-9a61-4278-82cc-962656385e73", 00:33:47.809 "assigned_rate_limits": { 00:33:47.809 "rw_ios_per_sec": 0, 00:33:47.809 "rw_mbytes_per_sec": 0, 00:33:47.809 "r_mbytes_per_sec": 0, 00:33:47.809 "w_mbytes_per_sec": 0 00:33:47.809 }, 00:33:47.809 "claimed": false, 00:33:47.809 "zoned": false, 00:33:47.809 "supported_io_types": { 00:33:47.809 "read": true, 00:33:47.809 "write": true, 00:33:47.809 "unmap": false, 00:33:47.809 "flush": false, 00:33:47.809 "reset": true, 00:33:47.809 "nvme_admin": false, 00:33:47.809 "nvme_io": false, 00:33:47.809 "nvme_io_md": false, 00:33:47.809 "write_zeroes": true, 00:33:47.809 "zcopy": false, 00:33:47.809 "get_zone_info": false, 00:33:47.809 "zone_management": false, 00:33:47.809 "zone_append": false, 00:33:47.809 "compare": false, 00:33:47.809 "compare_and_write": false, 00:33:47.809 "abort": false, 00:33:47.809 "seek_hole": false, 00:33:47.809 "seek_data": false, 00:33:47.809 "copy": false, 00:33:47.809 "nvme_iov_md": false 00:33:47.809 }, 00:33:47.809 "driver_specific": { 00:33:47.809 "raid": { 00:33:47.809 "uuid": "c63ab43f-9a61-4278-82cc-962656385e73", 00:33:47.809 "strip_size_kb": 64, 00:33:47.809 "state": "online", 00:33:47.809 "raid_level": "raid5f", 00:33:47.809 "superblock": true, 00:33:47.809 "num_base_bdevs": 4, 00:33:47.809 "num_base_bdevs_discovered": 4, 00:33:47.809 "num_base_bdevs_operational": 4, 00:33:47.809 "base_bdevs_list": [ 00:33:47.809 { 00:33:47.809 "name": "pt1", 00:33:47.809 "uuid": "00000000-0000-0000-0000-000000000001", 00:33:47.809 "is_configured": true, 00:33:47.809 "data_offset": 2048, 00:33:47.809 "data_size": 63488 00:33:47.809 }, 00:33:47.809 { 00:33:47.809 "name": "pt2", 00:33:47.809 "uuid": "00000000-0000-0000-0000-000000000002", 00:33:47.809 "is_configured": true, 00:33:47.809 "data_offset": 2048, 00:33:47.809 "data_size": 63488 00:33:47.809 }, 00:33:47.809 { 00:33:47.809 "name": "pt3", 00:33:47.809 "uuid": "00000000-0000-0000-0000-000000000003", 00:33:47.809 "is_configured": true, 00:33:47.809 "data_offset": 2048, 00:33:47.809 "data_size": 63488 00:33:47.809 }, 00:33:47.809 { 00:33:47.809 "name": "pt4", 00:33:47.809 "uuid": "00000000-0000-0000-0000-000000000004", 00:33:47.809 "is_configured": true, 00:33:47.809 "data_offset": 2048, 00:33:47.809 "data_size": 63488 00:33:47.809 } 00:33:47.809 ] 00:33:47.809 } 00:33:47.809 } 00:33:47.809 }' 00:33:47.809 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:33:47.809 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:33:47.809 pt2 00:33:47.809 pt3 00:33:47.809 pt4' 00:33:47.809 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:47.809 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:33:47.809 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:47.809 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:47.809 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:33:47.809 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:47.809 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:47.809 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:47.809 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:33:47.809 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:33:47.809 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:47.809 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:47.810 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:33:47.810 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:47.810 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:47.810 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:47.810 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:33:47.810 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:33:47.810 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:47.810 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:47.810 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:33:47.810 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:47.810 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:47.810 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:47.810 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:33:47.810 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:33:47.810 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:47.810 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:47.810 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:33:47.810 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:47.810 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:33:48.068 [2024-11-20 13:53:54.889460] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' c63ab43f-9a61-4278-82cc-962656385e73 '!=' c63ab43f-9a61-4278-82cc-962656385e73 ']' 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid5f 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:48.068 [2024-11-20 13:53:54.921327] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:48.068 "name": "raid_bdev1", 00:33:48.068 "uuid": "c63ab43f-9a61-4278-82cc-962656385e73", 00:33:48.068 "strip_size_kb": 64, 00:33:48.068 "state": "online", 00:33:48.068 "raid_level": "raid5f", 00:33:48.068 "superblock": true, 00:33:48.068 "num_base_bdevs": 4, 00:33:48.068 "num_base_bdevs_discovered": 3, 00:33:48.068 "num_base_bdevs_operational": 3, 00:33:48.068 "base_bdevs_list": [ 00:33:48.068 { 00:33:48.068 "name": null, 00:33:48.068 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:48.068 "is_configured": false, 00:33:48.068 "data_offset": 0, 00:33:48.068 "data_size": 63488 00:33:48.068 }, 00:33:48.068 { 00:33:48.068 "name": "pt2", 00:33:48.068 "uuid": "00000000-0000-0000-0000-000000000002", 00:33:48.068 "is_configured": true, 00:33:48.068 "data_offset": 2048, 00:33:48.068 "data_size": 63488 00:33:48.068 }, 00:33:48.068 { 00:33:48.068 "name": "pt3", 00:33:48.068 "uuid": "00000000-0000-0000-0000-000000000003", 00:33:48.068 "is_configured": true, 00:33:48.068 "data_offset": 2048, 00:33:48.068 "data_size": 63488 00:33:48.068 }, 00:33:48.068 { 00:33:48.068 "name": "pt4", 00:33:48.068 "uuid": "00000000-0000-0000-0000-000000000004", 00:33:48.068 "is_configured": true, 00:33:48.068 "data_offset": 2048, 00:33:48.068 "data_size": 63488 00:33:48.068 } 00:33:48.068 ] 00:33:48.068 }' 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:48.068 13:53:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:48.326 [2024-11-20 13:53:55.253358] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:33:48.326 [2024-11-20 13:53:55.253386] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:33:48.326 [2024-11-20 13:53:55.253464] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:33:48.326 [2024-11-20 13:53:55.253541] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:33:48.326 [2024-11-20 13:53:55.253549] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt4 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:48.326 [2024-11-20 13:53:55.325347] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:33:48.326 [2024-11-20 13:53:55.325482] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:48.326 [2024-11-20 13:53:55.325504] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:33:48.326 [2024-11-20 13:53:55.325512] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:48.326 [2024-11-20 13:53:55.327535] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:48.326 [2024-11-20 13:53:55.327568] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:33:48.326 [2024-11-20 13:53:55.327643] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:33:48.326 [2024-11-20 13:53:55.327683] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:33:48.326 pt2 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:48.326 "name": "raid_bdev1", 00:33:48.326 "uuid": "c63ab43f-9a61-4278-82cc-962656385e73", 00:33:48.326 "strip_size_kb": 64, 00:33:48.326 "state": "configuring", 00:33:48.326 "raid_level": "raid5f", 00:33:48.326 "superblock": true, 00:33:48.326 "num_base_bdevs": 4, 00:33:48.326 "num_base_bdevs_discovered": 1, 00:33:48.326 "num_base_bdevs_operational": 3, 00:33:48.326 "base_bdevs_list": [ 00:33:48.326 { 00:33:48.326 "name": null, 00:33:48.326 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:48.326 "is_configured": false, 00:33:48.326 "data_offset": 2048, 00:33:48.326 "data_size": 63488 00:33:48.326 }, 00:33:48.326 { 00:33:48.326 "name": "pt2", 00:33:48.326 "uuid": "00000000-0000-0000-0000-000000000002", 00:33:48.326 "is_configured": true, 00:33:48.326 "data_offset": 2048, 00:33:48.326 "data_size": 63488 00:33:48.326 }, 00:33:48.326 { 00:33:48.326 "name": null, 00:33:48.326 "uuid": "00000000-0000-0000-0000-000000000003", 00:33:48.326 "is_configured": false, 00:33:48.326 "data_offset": 2048, 00:33:48.326 "data_size": 63488 00:33:48.326 }, 00:33:48.326 { 00:33:48.326 "name": null, 00:33:48.326 "uuid": "00000000-0000-0000-0000-000000000004", 00:33:48.326 "is_configured": false, 00:33:48.326 "data_offset": 2048, 00:33:48.326 "data_size": 63488 00:33:48.326 } 00:33:48.326 ] 00:33:48.326 }' 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:48.326 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:48.909 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:33:48.909 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:33:48.909 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:33:48.909 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:48.909 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:48.909 [2024-11-20 13:53:55.649450] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:33:48.909 [2024-11-20 13:53:55.649531] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:48.909 [2024-11-20 13:53:55.649552] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:33:48.909 [2024-11-20 13:53:55.649561] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:48.909 [2024-11-20 13:53:55.650018] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:48.909 [2024-11-20 13:53:55.650038] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:33:48.909 [2024-11-20 13:53:55.650135] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:33:48.909 [2024-11-20 13:53:55.650158] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:33:48.909 pt3 00:33:48.909 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:48.909 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:33:48.909 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:48.909 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:48.909 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:48.909 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:48.909 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:33:48.909 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:48.909 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:48.909 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:48.909 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:48.909 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:48.909 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:48.909 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:48.909 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:48.909 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:48.909 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:48.909 "name": "raid_bdev1", 00:33:48.909 "uuid": "c63ab43f-9a61-4278-82cc-962656385e73", 00:33:48.909 "strip_size_kb": 64, 00:33:48.909 "state": "configuring", 00:33:48.909 "raid_level": "raid5f", 00:33:48.909 "superblock": true, 00:33:48.909 "num_base_bdevs": 4, 00:33:48.909 "num_base_bdevs_discovered": 2, 00:33:48.909 "num_base_bdevs_operational": 3, 00:33:48.909 "base_bdevs_list": [ 00:33:48.909 { 00:33:48.909 "name": null, 00:33:48.909 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:48.909 "is_configured": false, 00:33:48.909 "data_offset": 2048, 00:33:48.909 "data_size": 63488 00:33:48.909 }, 00:33:48.909 { 00:33:48.909 "name": "pt2", 00:33:48.909 "uuid": "00000000-0000-0000-0000-000000000002", 00:33:48.909 "is_configured": true, 00:33:48.909 "data_offset": 2048, 00:33:48.909 "data_size": 63488 00:33:48.909 }, 00:33:48.909 { 00:33:48.909 "name": "pt3", 00:33:48.909 "uuid": "00000000-0000-0000-0000-000000000003", 00:33:48.909 "is_configured": true, 00:33:48.909 "data_offset": 2048, 00:33:48.909 "data_size": 63488 00:33:48.909 }, 00:33:48.909 { 00:33:48.909 "name": null, 00:33:48.909 "uuid": "00000000-0000-0000-0000-000000000004", 00:33:48.909 "is_configured": false, 00:33:48.909 "data_offset": 2048, 00:33:48.909 "data_size": 63488 00:33:48.909 } 00:33:48.909 ] 00:33:48.909 }' 00:33:48.909 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:48.909 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:49.168 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:33:49.168 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:33:49.168 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@519 -- # i=3 00:33:49.168 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:33:49.168 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:49.168 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:49.168 [2024-11-20 13:53:55.973540] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:33:49.168 [2024-11-20 13:53:55.973619] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:49.168 [2024-11-20 13:53:55.973640] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:33:49.168 [2024-11-20 13:53:55.973649] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:49.168 [2024-11-20 13:53:55.974083] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:49.168 [2024-11-20 13:53:55.974098] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:33:49.168 [2024-11-20 13:53:55.974176] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:33:49.168 [2024-11-20 13:53:55.974201] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:33:49.168 [2024-11-20 13:53:55.974317] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:33:49.168 [2024-11-20 13:53:55.974325] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:33:49.168 [2024-11-20 13:53:55.974540] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:33:49.168 [2024-11-20 13:53:55.978414] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:33:49.168 [2024-11-20 13:53:55.978436] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:33:49.168 [2024-11-20 13:53:55.978687] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:49.168 pt4 00:33:49.168 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:49.168 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:33:49.168 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:49.168 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:49.168 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:49.168 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:49.168 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:33:49.168 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:49.168 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:49.168 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:49.168 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:49.168 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:49.168 13:53:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:49.168 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:49.168 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:49.168 13:53:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:49.168 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:49.168 "name": "raid_bdev1", 00:33:49.168 "uuid": "c63ab43f-9a61-4278-82cc-962656385e73", 00:33:49.168 "strip_size_kb": 64, 00:33:49.168 "state": "online", 00:33:49.168 "raid_level": "raid5f", 00:33:49.168 "superblock": true, 00:33:49.168 "num_base_bdevs": 4, 00:33:49.168 "num_base_bdevs_discovered": 3, 00:33:49.168 "num_base_bdevs_operational": 3, 00:33:49.168 "base_bdevs_list": [ 00:33:49.168 { 00:33:49.168 "name": null, 00:33:49.168 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:49.168 "is_configured": false, 00:33:49.168 "data_offset": 2048, 00:33:49.168 "data_size": 63488 00:33:49.168 }, 00:33:49.168 { 00:33:49.168 "name": "pt2", 00:33:49.168 "uuid": "00000000-0000-0000-0000-000000000002", 00:33:49.168 "is_configured": true, 00:33:49.168 "data_offset": 2048, 00:33:49.168 "data_size": 63488 00:33:49.168 }, 00:33:49.168 { 00:33:49.168 "name": "pt3", 00:33:49.168 "uuid": "00000000-0000-0000-0000-000000000003", 00:33:49.168 "is_configured": true, 00:33:49.168 "data_offset": 2048, 00:33:49.168 "data_size": 63488 00:33:49.168 }, 00:33:49.168 { 00:33:49.168 "name": "pt4", 00:33:49.168 "uuid": "00000000-0000-0000-0000-000000000004", 00:33:49.168 "is_configured": true, 00:33:49.168 "data_offset": 2048, 00:33:49.168 "data_size": 63488 00:33:49.168 } 00:33:49.168 ] 00:33:49.168 }' 00:33:49.168 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:49.168 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:49.427 [2024-11-20 13:53:56.283344] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:33:49.427 [2024-11-20 13:53:56.283379] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:33:49.427 [2024-11-20 13:53:56.283457] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:33:49.427 [2024-11-20 13:53:56.283533] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:33:49.427 [2024-11-20 13:53:56.283543] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 4 -gt 2 ']' 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@534 -- # i=3 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt4 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:49.427 [2024-11-20 13:53:56.331323] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:33:49.427 [2024-11-20 13:53:56.331517] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:49.427 [2024-11-20 13:53:56.331548] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c080 00:33:49.427 [2024-11-20 13:53:56.331563] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:49.427 [2024-11-20 13:53:56.333622] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:49.427 [2024-11-20 13:53:56.333653] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:33:49.427 [2024-11-20 13:53:56.333730] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:33:49.427 [2024-11-20 13:53:56.333773] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:33:49.427 [2024-11-20 13:53:56.333881] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:33:49.427 [2024-11-20 13:53:56.333892] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:33:49.427 [2024-11-20 13:53:56.333906] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:33:49.427 [2024-11-20 13:53:56.333960] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:33:49.427 [2024-11-20 13:53:56.334061] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:33:49.427 pt1 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 4 -gt 2 ']' 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:49.427 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:49.428 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:49.428 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:49.428 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:49.428 "name": "raid_bdev1", 00:33:49.428 "uuid": "c63ab43f-9a61-4278-82cc-962656385e73", 00:33:49.428 "strip_size_kb": 64, 00:33:49.428 "state": "configuring", 00:33:49.428 "raid_level": "raid5f", 00:33:49.428 "superblock": true, 00:33:49.428 "num_base_bdevs": 4, 00:33:49.428 "num_base_bdevs_discovered": 2, 00:33:49.428 "num_base_bdevs_operational": 3, 00:33:49.428 "base_bdevs_list": [ 00:33:49.428 { 00:33:49.428 "name": null, 00:33:49.428 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:49.428 "is_configured": false, 00:33:49.428 "data_offset": 2048, 00:33:49.428 "data_size": 63488 00:33:49.428 }, 00:33:49.428 { 00:33:49.428 "name": "pt2", 00:33:49.428 "uuid": "00000000-0000-0000-0000-000000000002", 00:33:49.428 "is_configured": true, 00:33:49.428 "data_offset": 2048, 00:33:49.428 "data_size": 63488 00:33:49.428 }, 00:33:49.428 { 00:33:49.428 "name": "pt3", 00:33:49.428 "uuid": "00000000-0000-0000-0000-000000000003", 00:33:49.428 "is_configured": true, 00:33:49.428 "data_offset": 2048, 00:33:49.428 "data_size": 63488 00:33:49.428 }, 00:33:49.428 { 00:33:49.428 "name": null, 00:33:49.428 "uuid": "00000000-0000-0000-0000-000000000004", 00:33:49.428 "is_configured": false, 00:33:49.428 "data_offset": 2048, 00:33:49.428 "data_size": 63488 00:33:49.428 } 00:33:49.428 ] 00:33:49.428 }' 00:33:49.428 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:49.428 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:49.686 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:33:49.686 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:33:49.686 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:49.686 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:49.686 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:49.686 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:33:49.686 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:33:49.686 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:49.686 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:49.686 [2024-11-20 13:53:56.707444] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:33:49.686 [2024-11-20 13:53:56.707522] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:49.686 [2024-11-20 13:53:56.707545] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c680 00:33:49.686 [2024-11-20 13:53:56.707554] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:49.686 [2024-11-20 13:53:56.707973] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:49.686 [2024-11-20 13:53:56.708008] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:33:49.686 [2024-11-20 13:53:56.708088] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:33:49.686 [2024-11-20 13:53:56.708108] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:33:49.686 [2024-11-20 13:53:56.708226] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:33:49.686 [2024-11-20 13:53:56.708238] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:33:49.686 [2024-11-20 13:53:56.708455] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:33:49.686 [2024-11-20 13:53:56.712399] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:33:49.686 [2024-11-20 13:53:56.712421] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:33:49.686 [2024-11-20 13:53:56.712678] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:49.686 pt4 00:33:49.686 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:49.687 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:33:49.687 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:49.687 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:49.687 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:49.687 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:49.687 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:33:49.687 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:49.687 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:49.687 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:49.687 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:49.687 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:49.687 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:49.687 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:49.687 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:49.687 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:49.945 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:49.945 "name": "raid_bdev1", 00:33:49.945 "uuid": "c63ab43f-9a61-4278-82cc-962656385e73", 00:33:49.945 "strip_size_kb": 64, 00:33:49.945 "state": "online", 00:33:49.945 "raid_level": "raid5f", 00:33:49.945 "superblock": true, 00:33:49.945 "num_base_bdevs": 4, 00:33:49.945 "num_base_bdevs_discovered": 3, 00:33:49.945 "num_base_bdevs_operational": 3, 00:33:49.945 "base_bdevs_list": [ 00:33:49.945 { 00:33:49.945 "name": null, 00:33:49.945 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:49.945 "is_configured": false, 00:33:49.945 "data_offset": 2048, 00:33:49.945 "data_size": 63488 00:33:49.945 }, 00:33:49.945 { 00:33:49.945 "name": "pt2", 00:33:49.945 "uuid": "00000000-0000-0000-0000-000000000002", 00:33:49.945 "is_configured": true, 00:33:49.945 "data_offset": 2048, 00:33:49.945 "data_size": 63488 00:33:49.945 }, 00:33:49.945 { 00:33:49.945 "name": "pt3", 00:33:49.945 "uuid": "00000000-0000-0000-0000-000000000003", 00:33:49.945 "is_configured": true, 00:33:49.945 "data_offset": 2048, 00:33:49.945 "data_size": 63488 00:33:49.945 }, 00:33:49.945 { 00:33:49.945 "name": "pt4", 00:33:49.945 "uuid": "00000000-0000-0000-0000-000000000004", 00:33:49.945 "is_configured": true, 00:33:49.945 "data_offset": 2048, 00:33:49.945 "data_size": 63488 00:33:49.945 } 00:33:49.945 ] 00:33:49.945 }' 00:33:49.945 13:53:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:49.945 13:53:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:50.203 13:53:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:33:50.203 13:53:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:33:50.203 13:53:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:50.203 13:53:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:50.203 13:53:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:50.203 13:53:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:33:50.203 13:53:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:33:50.204 13:53:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:33:50.204 13:53:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:50.204 13:53:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:50.204 [2024-11-20 13:53:57.065399] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:33:50.204 13:53:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:50.204 13:53:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' c63ab43f-9a61-4278-82cc-962656385e73 '!=' c63ab43f-9a61-4278-82cc-962656385e73 ']' 00:33:50.204 13:53:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 81656 00:33:50.204 13:53:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 81656 ']' 00:33:50.204 13:53:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@958 -- # kill -0 81656 00:33:50.204 13:53:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # uname 00:33:50.204 13:53:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:33:50.204 13:53:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 81656 00:33:50.204 killing process with pid 81656 00:33:50.204 13:53:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:33:50.204 13:53:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:33:50.204 13:53:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 81656' 00:33:50.204 13:53:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@973 -- # kill 81656 00:33:50.204 13:53:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@978 -- # wait 81656 00:33:50.204 [2024-11-20 13:53:57.117592] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:33:50.204 [2024-11-20 13:53:57.117701] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:33:50.204 [2024-11-20 13:53:57.117779] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:33:50.204 [2024-11-20 13:53:57.117796] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:33:50.462 [2024-11-20 13:53:57.327718] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:33:51.033 ************************************ 00:33:51.033 END TEST raid5f_superblock_test 00:33:51.033 ************************************ 00:33:51.033 13:53:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:33:51.033 00:33:51.033 real 0m6.080s 00:33:51.033 user 0m9.589s 00:33:51.033 sys 0m1.067s 00:33:51.033 13:53:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:33:51.033 13:53:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:33:51.033 13:53:57 bdev_raid -- bdev/bdev_raid.sh@989 -- # '[' true = true ']' 00:33:51.033 13:53:57 bdev_raid -- bdev/bdev_raid.sh@990 -- # run_test raid5f_rebuild_test raid_rebuild_test raid5f 4 false false true 00:33:51.033 13:53:57 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:33:51.033 13:53:57 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:33:51.033 13:53:57 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:33:51.033 ************************************ 00:33:51.033 START TEST raid5f_rebuild_test 00:33:51.033 ************************************ 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 4 false false true 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=82119 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 82119 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 82119 ']' 00:33:51.033 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:51.033 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:33:51.033 I/O size of 3145728 is greater than zero copy threshold (65536). 00:33:51.033 Zero copy mechanism will not be used. 00:33:51.033 [2024-11-20 13:53:58.078017] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:33:51.033 [2024-11-20 13:53:58.078135] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid82119 ] 00:33:51.293 [2024-11-20 13:53:58.232913] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:33:51.293 [2024-11-20 13:53:58.335053] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:33:51.552 [2024-11-20 13:53:58.459888] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:33:51.552 [2024-11-20 13:53:58.459945] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:33:52.119 13:53:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:33:52.119 13:53:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:33:52.119 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:33:52.119 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:33:52.119 13:53:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:52.119 13:53:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:52.119 BaseBdev1_malloc 00:33:52.119 13:53:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:52.119 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:33:52.119 13:53:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:52.119 13:53:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:52.119 [2024-11-20 13:53:58.951546] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:33:52.119 [2024-11-20 13:53:58.951774] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:52.119 [2024-11-20 13:53:58.951801] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:33:52.119 [2024-11-20 13:53:58.951812] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:52.119 [2024-11-20 13:53:58.953788] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:52.119 [2024-11-20 13:53:58.953826] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:33:52.119 BaseBdev1 00:33:52.119 13:53:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:52.119 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:33:52.119 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:33:52.119 13:53:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:52.119 13:53:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:52.119 BaseBdev2_malloc 00:33:52.119 13:53:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:52.119 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:33:52.119 13:53:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:52.119 13:53:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:52.119 [2024-11-20 13:53:58.985235] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:33:52.119 [2024-11-20 13:53:58.985289] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:52.119 [2024-11-20 13:53:58.985307] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:33:52.119 [2024-11-20 13:53:58.985317] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:52.119 [2024-11-20 13:53:58.987222] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:52.119 [2024-11-20 13:53:58.987253] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:33:52.119 BaseBdev2 00:33:52.119 13:53:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:52.119 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:33:52.119 13:53:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:33:52.119 13:53:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:52.119 13:53:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:52.119 BaseBdev3_malloc 00:33:52.119 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:52.119 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:33:52.119 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:52.119 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:52.119 [2024-11-20 13:53:59.041433] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:33:52.119 [2024-11-20 13:53:59.041664] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:52.119 [2024-11-20 13:53:59.041693] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:33:52.119 [2024-11-20 13:53:59.041703] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:52.119 [2024-11-20 13:53:59.043600] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:52.119 [2024-11-20 13:53:59.043634] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:33:52.119 BaseBdev3 00:33:52.119 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:52.119 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:33:52.119 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:33:52.119 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:52.119 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:52.119 BaseBdev4_malloc 00:33:52.119 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:52.119 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:33:52.119 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:52.119 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:52.119 [2024-11-20 13:53:59.079179] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:33:52.119 [2024-11-20 13:53:59.079235] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:52.119 [2024-11-20 13:53:59.079251] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:33:52.119 [2024-11-20 13:53:59.079262] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:52.119 [2024-11-20 13:53:59.081175] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:52.119 [2024-11-20 13:53:59.081210] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:33:52.119 BaseBdev4 00:33:52.119 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:52.119 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:33:52.119 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:52.119 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:52.119 spare_malloc 00:33:52.119 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:52.120 spare_delay 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:52.120 [2024-11-20 13:53:59.124639] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:33:52.120 [2024-11-20 13:53:59.124694] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:52.120 [2024-11-20 13:53:59.124709] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:33:52.120 [2024-11-20 13:53:59.124719] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:52.120 [2024-11-20 13:53:59.126587] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:52.120 [2024-11-20 13:53:59.126619] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:33:52.120 spare 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:52.120 [2024-11-20 13:53:59.132673] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:33:52.120 [2024-11-20 13:53:59.134288] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:33:52.120 [2024-11-20 13:53:59.134341] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:33:52.120 [2024-11-20 13:53:59.134383] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:33:52.120 [2024-11-20 13:53:59.134453] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:33:52.120 [2024-11-20 13:53:59.134464] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:33:52.120 [2024-11-20 13:53:59.134686] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:33:52.120 [2024-11-20 13:53:59.138721] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:33:52.120 [2024-11-20 13:53:59.138739] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:33:52.120 [2024-11-20 13:53:59.138897] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:52.120 "name": "raid_bdev1", 00:33:52.120 "uuid": "ff199959-08c2-4ac2-a4f9-969e7dd3ac3e", 00:33:52.120 "strip_size_kb": 64, 00:33:52.120 "state": "online", 00:33:52.120 "raid_level": "raid5f", 00:33:52.120 "superblock": false, 00:33:52.120 "num_base_bdevs": 4, 00:33:52.120 "num_base_bdevs_discovered": 4, 00:33:52.120 "num_base_bdevs_operational": 4, 00:33:52.120 "base_bdevs_list": [ 00:33:52.120 { 00:33:52.120 "name": "BaseBdev1", 00:33:52.120 "uuid": "bed71a62-4697-5012-8f04-3bbd3d915f86", 00:33:52.120 "is_configured": true, 00:33:52.120 "data_offset": 0, 00:33:52.120 "data_size": 65536 00:33:52.120 }, 00:33:52.120 { 00:33:52.120 "name": "BaseBdev2", 00:33:52.120 "uuid": "ef46d29f-fb56-5348-9e49-4a3825c02f93", 00:33:52.120 "is_configured": true, 00:33:52.120 "data_offset": 0, 00:33:52.120 "data_size": 65536 00:33:52.120 }, 00:33:52.120 { 00:33:52.120 "name": "BaseBdev3", 00:33:52.120 "uuid": "f2e1829c-1bf1-5ad2-845b-101076e9f258", 00:33:52.120 "is_configured": true, 00:33:52.120 "data_offset": 0, 00:33:52.120 "data_size": 65536 00:33:52.120 }, 00:33:52.120 { 00:33:52.120 "name": "BaseBdev4", 00:33:52.120 "uuid": "f3e91a18-a294-5dc8-96a6-06657f9af8da", 00:33:52.120 "is_configured": true, 00:33:52.120 "data_offset": 0, 00:33:52.120 "data_size": 65536 00:33:52.120 } 00:33:52.120 ] 00:33:52.120 }' 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:52.120 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:52.686 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:33:52.686 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:33:52.686 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:52.686 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:52.686 [2024-11-20 13:53:59.459848] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:33:52.686 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:52.686 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=196608 00:33:52.686 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:33:52.686 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:52.686 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:52.686 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:52.686 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:52.686 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:33:52.686 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:33:52.686 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:33:52.686 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:33:52.686 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:33:52.686 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:33:52.686 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:33:52.686 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:33:52.686 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:33:52.686 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:33:52.686 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:33:52.686 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:33:52.686 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:33:52.686 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:33:52.686 [2024-11-20 13:53:59.711742] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:33:52.686 /dev/nbd0 00:33:52.943 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:33:52.943 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:33:52.943 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:33:52.943 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:33:52.944 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:33:52.944 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:33:52.944 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:33:52.944 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:33:52.944 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:33:52.944 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:33:52.944 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:33:52.944 1+0 records in 00:33:52.944 1+0 records out 00:33:52.944 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000210111 s, 19.5 MB/s 00:33:52.944 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:33:52.944 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:33:52.944 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:33:52.944 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:33:52.944 13:53:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:33:52.944 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:33:52.944 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:33:52.944 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:33:52.944 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@630 -- # write_unit_size=384 00:33:52.944 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@631 -- # echo 192 00:33:52.944 13:53:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=196608 count=512 oflag=direct 00:33:53.509 512+0 records in 00:33:53.509 512+0 records out 00:33:53.509 100663296 bytes (101 MB, 96 MiB) copied, 0.498071 s, 202 MB/s 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:33:53.509 [2024-11-20 13:54:00.478187] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:53.509 [2024-11-20 13:54:00.486995] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:53.509 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:53.509 "name": "raid_bdev1", 00:33:53.509 "uuid": "ff199959-08c2-4ac2-a4f9-969e7dd3ac3e", 00:33:53.509 "strip_size_kb": 64, 00:33:53.509 "state": "online", 00:33:53.509 "raid_level": "raid5f", 00:33:53.509 "superblock": false, 00:33:53.509 "num_base_bdevs": 4, 00:33:53.509 "num_base_bdevs_discovered": 3, 00:33:53.509 "num_base_bdevs_operational": 3, 00:33:53.509 "base_bdevs_list": [ 00:33:53.509 { 00:33:53.509 "name": null, 00:33:53.509 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:53.509 "is_configured": false, 00:33:53.509 "data_offset": 0, 00:33:53.509 "data_size": 65536 00:33:53.509 }, 00:33:53.509 { 00:33:53.509 "name": "BaseBdev2", 00:33:53.509 "uuid": "ef46d29f-fb56-5348-9e49-4a3825c02f93", 00:33:53.509 "is_configured": true, 00:33:53.509 "data_offset": 0, 00:33:53.509 "data_size": 65536 00:33:53.509 }, 00:33:53.509 { 00:33:53.509 "name": "BaseBdev3", 00:33:53.509 "uuid": "f2e1829c-1bf1-5ad2-845b-101076e9f258", 00:33:53.509 "is_configured": true, 00:33:53.509 "data_offset": 0, 00:33:53.509 "data_size": 65536 00:33:53.509 }, 00:33:53.509 { 00:33:53.509 "name": "BaseBdev4", 00:33:53.509 "uuid": "f3e91a18-a294-5dc8-96a6-06657f9af8da", 00:33:53.509 "is_configured": true, 00:33:53.509 "data_offset": 0, 00:33:53.509 "data_size": 65536 00:33:53.509 } 00:33:53.509 ] 00:33:53.509 }' 00:33:53.510 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:53.510 13:54:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:53.768 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:33:53.768 13:54:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:53.768 13:54:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:53.768 [2024-11-20 13:54:00.819068] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:33:54.025 [2024-11-20 13:54:00.827493] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b750 00:33:54.025 13:54:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:54.025 13:54:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:33:54.025 [2024-11-20 13:54:00.833158] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:33:54.959 13:54:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:54.959 13:54:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:54.959 13:54:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:54.959 13:54:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:54.959 13:54:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:54.959 13:54:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:54.959 13:54:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:54.959 13:54:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:54.959 13:54:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:54.959 13:54:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:54.959 13:54:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:54.959 "name": "raid_bdev1", 00:33:54.959 "uuid": "ff199959-08c2-4ac2-a4f9-969e7dd3ac3e", 00:33:54.959 "strip_size_kb": 64, 00:33:54.959 "state": "online", 00:33:54.959 "raid_level": "raid5f", 00:33:54.959 "superblock": false, 00:33:54.959 "num_base_bdevs": 4, 00:33:54.959 "num_base_bdevs_discovered": 4, 00:33:54.959 "num_base_bdevs_operational": 4, 00:33:54.959 "process": { 00:33:54.959 "type": "rebuild", 00:33:54.959 "target": "spare", 00:33:54.959 "progress": { 00:33:54.959 "blocks": 19200, 00:33:54.959 "percent": 9 00:33:54.959 } 00:33:54.959 }, 00:33:54.959 "base_bdevs_list": [ 00:33:54.959 { 00:33:54.959 "name": "spare", 00:33:54.959 "uuid": "78be2159-271b-5a17-bd0d-4657ec9e8724", 00:33:54.959 "is_configured": true, 00:33:54.959 "data_offset": 0, 00:33:54.959 "data_size": 65536 00:33:54.959 }, 00:33:54.959 { 00:33:54.959 "name": "BaseBdev2", 00:33:54.959 "uuid": "ef46d29f-fb56-5348-9e49-4a3825c02f93", 00:33:54.959 "is_configured": true, 00:33:54.959 "data_offset": 0, 00:33:54.959 "data_size": 65536 00:33:54.959 }, 00:33:54.959 { 00:33:54.959 "name": "BaseBdev3", 00:33:54.959 "uuid": "f2e1829c-1bf1-5ad2-845b-101076e9f258", 00:33:54.959 "is_configured": true, 00:33:54.959 "data_offset": 0, 00:33:54.959 "data_size": 65536 00:33:54.959 }, 00:33:54.959 { 00:33:54.959 "name": "BaseBdev4", 00:33:54.959 "uuid": "f3e91a18-a294-5dc8-96a6-06657f9af8da", 00:33:54.959 "is_configured": true, 00:33:54.959 "data_offset": 0, 00:33:54.959 "data_size": 65536 00:33:54.959 } 00:33:54.959 ] 00:33:54.959 }' 00:33:54.959 13:54:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:54.959 13:54:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:33:54.959 13:54:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:54.959 13:54:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:33:54.959 13:54:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:33:54.959 13:54:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:54.959 13:54:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:54.959 [2024-11-20 13:54:01.942844] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:33:55.217 [2024-11-20 13:54:02.043203] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:33:55.217 [2024-11-20 13:54:02.043278] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:55.217 [2024-11-20 13:54:02.043295] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:33:55.217 [2024-11-20 13:54:02.043307] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:33:55.217 13:54:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:55.217 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:33:55.217 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:55.217 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:55.217 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:33:55.217 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:33:55.217 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:33:55.217 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:55.217 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:55.217 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:55.217 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:55.217 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:55.217 13:54:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:55.217 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:55.217 13:54:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:55.217 13:54:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:55.217 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:55.217 "name": "raid_bdev1", 00:33:55.217 "uuid": "ff199959-08c2-4ac2-a4f9-969e7dd3ac3e", 00:33:55.217 "strip_size_kb": 64, 00:33:55.217 "state": "online", 00:33:55.217 "raid_level": "raid5f", 00:33:55.217 "superblock": false, 00:33:55.217 "num_base_bdevs": 4, 00:33:55.217 "num_base_bdevs_discovered": 3, 00:33:55.217 "num_base_bdevs_operational": 3, 00:33:55.217 "base_bdevs_list": [ 00:33:55.217 { 00:33:55.217 "name": null, 00:33:55.217 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:55.217 "is_configured": false, 00:33:55.217 "data_offset": 0, 00:33:55.217 "data_size": 65536 00:33:55.217 }, 00:33:55.217 { 00:33:55.217 "name": "BaseBdev2", 00:33:55.217 "uuid": "ef46d29f-fb56-5348-9e49-4a3825c02f93", 00:33:55.217 "is_configured": true, 00:33:55.217 "data_offset": 0, 00:33:55.217 "data_size": 65536 00:33:55.217 }, 00:33:55.217 { 00:33:55.217 "name": "BaseBdev3", 00:33:55.217 "uuid": "f2e1829c-1bf1-5ad2-845b-101076e9f258", 00:33:55.217 "is_configured": true, 00:33:55.217 "data_offset": 0, 00:33:55.217 "data_size": 65536 00:33:55.217 }, 00:33:55.217 { 00:33:55.217 "name": "BaseBdev4", 00:33:55.217 "uuid": "f3e91a18-a294-5dc8-96a6-06657f9af8da", 00:33:55.217 "is_configured": true, 00:33:55.217 "data_offset": 0, 00:33:55.217 "data_size": 65536 00:33:55.217 } 00:33:55.217 ] 00:33:55.217 }' 00:33:55.217 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:55.217 13:54:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:55.475 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:33:55.475 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:55.475 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:33:55.475 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:33:55.475 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:55.475 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:55.475 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:55.475 13:54:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:55.475 13:54:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:55.475 13:54:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:55.476 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:55.476 "name": "raid_bdev1", 00:33:55.476 "uuid": "ff199959-08c2-4ac2-a4f9-969e7dd3ac3e", 00:33:55.476 "strip_size_kb": 64, 00:33:55.476 "state": "online", 00:33:55.476 "raid_level": "raid5f", 00:33:55.476 "superblock": false, 00:33:55.476 "num_base_bdevs": 4, 00:33:55.476 "num_base_bdevs_discovered": 3, 00:33:55.476 "num_base_bdevs_operational": 3, 00:33:55.476 "base_bdevs_list": [ 00:33:55.476 { 00:33:55.476 "name": null, 00:33:55.476 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:55.476 "is_configured": false, 00:33:55.476 "data_offset": 0, 00:33:55.476 "data_size": 65536 00:33:55.476 }, 00:33:55.476 { 00:33:55.476 "name": "BaseBdev2", 00:33:55.476 "uuid": "ef46d29f-fb56-5348-9e49-4a3825c02f93", 00:33:55.476 "is_configured": true, 00:33:55.476 "data_offset": 0, 00:33:55.476 "data_size": 65536 00:33:55.476 }, 00:33:55.476 { 00:33:55.476 "name": "BaseBdev3", 00:33:55.476 "uuid": "f2e1829c-1bf1-5ad2-845b-101076e9f258", 00:33:55.476 "is_configured": true, 00:33:55.476 "data_offset": 0, 00:33:55.476 "data_size": 65536 00:33:55.476 }, 00:33:55.476 { 00:33:55.476 "name": "BaseBdev4", 00:33:55.476 "uuid": "f3e91a18-a294-5dc8-96a6-06657f9af8da", 00:33:55.476 "is_configured": true, 00:33:55.476 "data_offset": 0, 00:33:55.476 "data_size": 65536 00:33:55.476 } 00:33:55.476 ] 00:33:55.476 }' 00:33:55.476 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:55.476 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:33:55.476 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:55.476 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:33:55.476 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:33:55.476 13:54:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:55.476 13:54:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:55.476 [2024-11-20 13:54:02.448800] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:33:55.476 [2024-11-20 13:54:02.456829] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b820 00:33:55.476 13:54:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:55.476 13:54:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:33:55.476 [2024-11-20 13:54:02.462369] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:33:56.410 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:56.410 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:56.410 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:56.410 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:56.410 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:56.410 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:56.410 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:56.410 13:54:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:56.410 13:54:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:56.669 "name": "raid_bdev1", 00:33:56.669 "uuid": "ff199959-08c2-4ac2-a4f9-969e7dd3ac3e", 00:33:56.669 "strip_size_kb": 64, 00:33:56.669 "state": "online", 00:33:56.669 "raid_level": "raid5f", 00:33:56.669 "superblock": false, 00:33:56.669 "num_base_bdevs": 4, 00:33:56.669 "num_base_bdevs_discovered": 4, 00:33:56.669 "num_base_bdevs_operational": 4, 00:33:56.669 "process": { 00:33:56.669 "type": "rebuild", 00:33:56.669 "target": "spare", 00:33:56.669 "progress": { 00:33:56.669 "blocks": 19200, 00:33:56.669 "percent": 9 00:33:56.669 } 00:33:56.669 }, 00:33:56.669 "base_bdevs_list": [ 00:33:56.669 { 00:33:56.669 "name": "spare", 00:33:56.669 "uuid": "78be2159-271b-5a17-bd0d-4657ec9e8724", 00:33:56.669 "is_configured": true, 00:33:56.669 "data_offset": 0, 00:33:56.669 "data_size": 65536 00:33:56.669 }, 00:33:56.669 { 00:33:56.669 "name": "BaseBdev2", 00:33:56.669 "uuid": "ef46d29f-fb56-5348-9e49-4a3825c02f93", 00:33:56.669 "is_configured": true, 00:33:56.669 "data_offset": 0, 00:33:56.669 "data_size": 65536 00:33:56.669 }, 00:33:56.669 { 00:33:56.669 "name": "BaseBdev3", 00:33:56.669 "uuid": "f2e1829c-1bf1-5ad2-845b-101076e9f258", 00:33:56.669 "is_configured": true, 00:33:56.669 "data_offset": 0, 00:33:56.669 "data_size": 65536 00:33:56.669 }, 00:33:56.669 { 00:33:56.669 "name": "BaseBdev4", 00:33:56.669 "uuid": "f3e91a18-a294-5dc8-96a6-06657f9af8da", 00:33:56.669 "is_configured": true, 00:33:56.669 "data_offset": 0, 00:33:56.669 "data_size": 65536 00:33:56.669 } 00:33:56.669 ] 00:33:56.669 }' 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=490 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:56.669 "name": "raid_bdev1", 00:33:56.669 "uuid": "ff199959-08c2-4ac2-a4f9-969e7dd3ac3e", 00:33:56.669 "strip_size_kb": 64, 00:33:56.669 "state": "online", 00:33:56.669 "raid_level": "raid5f", 00:33:56.669 "superblock": false, 00:33:56.669 "num_base_bdevs": 4, 00:33:56.669 "num_base_bdevs_discovered": 4, 00:33:56.669 "num_base_bdevs_operational": 4, 00:33:56.669 "process": { 00:33:56.669 "type": "rebuild", 00:33:56.669 "target": "spare", 00:33:56.669 "progress": { 00:33:56.669 "blocks": 21120, 00:33:56.669 "percent": 10 00:33:56.669 } 00:33:56.669 }, 00:33:56.669 "base_bdevs_list": [ 00:33:56.669 { 00:33:56.669 "name": "spare", 00:33:56.669 "uuid": "78be2159-271b-5a17-bd0d-4657ec9e8724", 00:33:56.669 "is_configured": true, 00:33:56.669 "data_offset": 0, 00:33:56.669 "data_size": 65536 00:33:56.669 }, 00:33:56.669 { 00:33:56.669 "name": "BaseBdev2", 00:33:56.669 "uuid": "ef46d29f-fb56-5348-9e49-4a3825c02f93", 00:33:56.669 "is_configured": true, 00:33:56.669 "data_offset": 0, 00:33:56.669 "data_size": 65536 00:33:56.669 }, 00:33:56.669 { 00:33:56.669 "name": "BaseBdev3", 00:33:56.669 "uuid": "f2e1829c-1bf1-5ad2-845b-101076e9f258", 00:33:56.669 "is_configured": true, 00:33:56.669 "data_offset": 0, 00:33:56.669 "data_size": 65536 00:33:56.669 }, 00:33:56.669 { 00:33:56.669 "name": "BaseBdev4", 00:33:56.669 "uuid": "f3e91a18-a294-5dc8-96a6-06657f9af8da", 00:33:56.669 "is_configured": true, 00:33:56.669 "data_offset": 0, 00:33:56.669 "data_size": 65536 00:33:56.669 } 00:33:56.669 ] 00:33:56.669 }' 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:33:56.669 13:54:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:33:58.043 13:54:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:33:58.043 13:54:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:58.043 13:54:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:58.043 13:54:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:58.043 13:54:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:58.043 13:54:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:58.043 13:54:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:58.043 13:54:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:58.044 13:54:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:58.044 13:54:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:58.044 13:54:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:58.044 13:54:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:58.044 "name": "raid_bdev1", 00:33:58.044 "uuid": "ff199959-08c2-4ac2-a4f9-969e7dd3ac3e", 00:33:58.044 "strip_size_kb": 64, 00:33:58.044 "state": "online", 00:33:58.044 "raid_level": "raid5f", 00:33:58.044 "superblock": false, 00:33:58.044 "num_base_bdevs": 4, 00:33:58.044 "num_base_bdevs_discovered": 4, 00:33:58.044 "num_base_bdevs_operational": 4, 00:33:58.044 "process": { 00:33:58.044 "type": "rebuild", 00:33:58.044 "target": "spare", 00:33:58.044 "progress": { 00:33:58.044 "blocks": 42240, 00:33:58.044 "percent": 21 00:33:58.044 } 00:33:58.044 }, 00:33:58.044 "base_bdevs_list": [ 00:33:58.044 { 00:33:58.044 "name": "spare", 00:33:58.044 "uuid": "78be2159-271b-5a17-bd0d-4657ec9e8724", 00:33:58.044 "is_configured": true, 00:33:58.044 "data_offset": 0, 00:33:58.044 "data_size": 65536 00:33:58.044 }, 00:33:58.044 { 00:33:58.044 "name": "BaseBdev2", 00:33:58.044 "uuid": "ef46d29f-fb56-5348-9e49-4a3825c02f93", 00:33:58.044 "is_configured": true, 00:33:58.044 "data_offset": 0, 00:33:58.044 "data_size": 65536 00:33:58.044 }, 00:33:58.044 { 00:33:58.044 "name": "BaseBdev3", 00:33:58.044 "uuid": "f2e1829c-1bf1-5ad2-845b-101076e9f258", 00:33:58.044 "is_configured": true, 00:33:58.044 "data_offset": 0, 00:33:58.044 "data_size": 65536 00:33:58.044 }, 00:33:58.044 { 00:33:58.044 "name": "BaseBdev4", 00:33:58.044 "uuid": "f3e91a18-a294-5dc8-96a6-06657f9af8da", 00:33:58.044 "is_configured": true, 00:33:58.044 "data_offset": 0, 00:33:58.044 "data_size": 65536 00:33:58.044 } 00:33:58.044 ] 00:33:58.044 }' 00:33:58.044 13:54:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:58.044 13:54:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:33:58.044 13:54:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:58.044 13:54:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:33:58.044 13:54:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:33:58.988 13:54:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:33:58.988 13:54:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:58.988 13:54:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:58.988 13:54:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:58.988 13:54:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:58.988 13:54:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:58.988 13:54:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:58.988 13:54:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:58.988 13:54:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:58.988 13:54:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:58.988 13:54:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:58.988 13:54:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:58.988 "name": "raid_bdev1", 00:33:58.988 "uuid": "ff199959-08c2-4ac2-a4f9-969e7dd3ac3e", 00:33:58.988 "strip_size_kb": 64, 00:33:58.988 "state": "online", 00:33:58.988 "raid_level": "raid5f", 00:33:58.988 "superblock": false, 00:33:58.988 "num_base_bdevs": 4, 00:33:58.988 "num_base_bdevs_discovered": 4, 00:33:58.988 "num_base_bdevs_operational": 4, 00:33:58.988 "process": { 00:33:58.988 "type": "rebuild", 00:33:58.988 "target": "spare", 00:33:58.988 "progress": { 00:33:58.988 "blocks": 63360, 00:33:58.988 "percent": 32 00:33:58.988 } 00:33:58.988 }, 00:33:58.988 "base_bdevs_list": [ 00:33:58.988 { 00:33:58.988 "name": "spare", 00:33:58.988 "uuid": "78be2159-271b-5a17-bd0d-4657ec9e8724", 00:33:58.988 "is_configured": true, 00:33:58.988 "data_offset": 0, 00:33:58.988 "data_size": 65536 00:33:58.988 }, 00:33:58.988 { 00:33:58.988 "name": "BaseBdev2", 00:33:58.988 "uuid": "ef46d29f-fb56-5348-9e49-4a3825c02f93", 00:33:58.988 "is_configured": true, 00:33:58.988 "data_offset": 0, 00:33:58.988 "data_size": 65536 00:33:58.988 }, 00:33:58.988 { 00:33:58.988 "name": "BaseBdev3", 00:33:58.988 "uuid": "f2e1829c-1bf1-5ad2-845b-101076e9f258", 00:33:58.988 "is_configured": true, 00:33:58.988 "data_offset": 0, 00:33:58.988 "data_size": 65536 00:33:58.988 }, 00:33:58.988 { 00:33:58.988 "name": "BaseBdev4", 00:33:58.988 "uuid": "f3e91a18-a294-5dc8-96a6-06657f9af8da", 00:33:58.988 "is_configured": true, 00:33:58.988 "data_offset": 0, 00:33:58.988 "data_size": 65536 00:33:58.988 } 00:33:58.988 ] 00:33:58.988 }' 00:33:58.988 13:54:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:58.988 13:54:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:33:58.988 13:54:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:58.988 13:54:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:33:58.988 13:54:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:33:59.922 13:54:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:33:59.923 13:54:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:59.923 13:54:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:59.923 13:54:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:59.923 13:54:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:59.923 13:54:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:59.923 13:54:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:59.923 13:54:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:59.923 13:54:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:33:59.923 13:54:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:33:59.923 13:54:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:33:59.923 13:54:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:59.923 "name": "raid_bdev1", 00:33:59.923 "uuid": "ff199959-08c2-4ac2-a4f9-969e7dd3ac3e", 00:33:59.923 "strip_size_kb": 64, 00:33:59.923 "state": "online", 00:33:59.923 "raid_level": "raid5f", 00:33:59.923 "superblock": false, 00:33:59.923 "num_base_bdevs": 4, 00:33:59.923 "num_base_bdevs_discovered": 4, 00:33:59.923 "num_base_bdevs_operational": 4, 00:33:59.923 "process": { 00:33:59.923 "type": "rebuild", 00:33:59.923 "target": "spare", 00:33:59.923 "progress": { 00:33:59.923 "blocks": 84480, 00:33:59.923 "percent": 42 00:33:59.923 } 00:33:59.923 }, 00:33:59.923 "base_bdevs_list": [ 00:33:59.923 { 00:33:59.923 "name": "spare", 00:33:59.923 "uuid": "78be2159-271b-5a17-bd0d-4657ec9e8724", 00:33:59.923 "is_configured": true, 00:33:59.923 "data_offset": 0, 00:33:59.923 "data_size": 65536 00:33:59.923 }, 00:33:59.923 { 00:33:59.923 "name": "BaseBdev2", 00:33:59.923 "uuid": "ef46d29f-fb56-5348-9e49-4a3825c02f93", 00:33:59.923 "is_configured": true, 00:33:59.923 "data_offset": 0, 00:33:59.923 "data_size": 65536 00:33:59.923 }, 00:33:59.923 { 00:33:59.923 "name": "BaseBdev3", 00:33:59.923 "uuid": "f2e1829c-1bf1-5ad2-845b-101076e9f258", 00:33:59.923 "is_configured": true, 00:33:59.923 "data_offset": 0, 00:33:59.923 "data_size": 65536 00:33:59.923 }, 00:33:59.923 { 00:33:59.923 "name": "BaseBdev4", 00:33:59.923 "uuid": "f3e91a18-a294-5dc8-96a6-06657f9af8da", 00:33:59.923 "is_configured": true, 00:33:59.923 "data_offset": 0, 00:33:59.923 "data_size": 65536 00:33:59.923 } 00:33:59.923 ] 00:33:59.923 }' 00:33:59.923 13:54:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:59.923 13:54:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:33:59.923 13:54:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:00.180 13:54:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:34:00.180 13:54:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:34:01.113 13:54:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:34:01.113 13:54:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:01.113 13:54:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:01.113 13:54:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:01.113 13:54:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:01.113 13:54:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:01.113 13:54:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:01.113 13:54:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:01.113 13:54:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:34:01.113 13:54:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:01.113 13:54:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:01.113 13:54:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:01.113 "name": "raid_bdev1", 00:34:01.113 "uuid": "ff199959-08c2-4ac2-a4f9-969e7dd3ac3e", 00:34:01.113 "strip_size_kb": 64, 00:34:01.113 "state": "online", 00:34:01.113 "raid_level": "raid5f", 00:34:01.113 "superblock": false, 00:34:01.113 "num_base_bdevs": 4, 00:34:01.113 "num_base_bdevs_discovered": 4, 00:34:01.113 "num_base_bdevs_operational": 4, 00:34:01.113 "process": { 00:34:01.113 "type": "rebuild", 00:34:01.113 "target": "spare", 00:34:01.113 "progress": { 00:34:01.113 "blocks": 105600, 00:34:01.113 "percent": 53 00:34:01.113 } 00:34:01.113 }, 00:34:01.113 "base_bdevs_list": [ 00:34:01.113 { 00:34:01.113 "name": "spare", 00:34:01.113 "uuid": "78be2159-271b-5a17-bd0d-4657ec9e8724", 00:34:01.113 "is_configured": true, 00:34:01.113 "data_offset": 0, 00:34:01.113 "data_size": 65536 00:34:01.113 }, 00:34:01.113 { 00:34:01.113 "name": "BaseBdev2", 00:34:01.113 "uuid": "ef46d29f-fb56-5348-9e49-4a3825c02f93", 00:34:01.113 "is_configured": true, 00:34:01.113 "data_offset": 0, 00:34:01.113 "data_size": 65536 00:34:01.113 }, 00:34:01.113 { 00:34:01.113 "name": "BaseBdev3", 00:34:01.113 "uuid": "f2e1829c-1bf1-5ad2-845b-101076e9f258", 00:34:01.113 "is_configured": true, 00:34:01.113 "data_offset": 0, 00:34:01.113 "data_size": 65536 00:34:01.113 }, 00:34:01.113 { 00:34:01.113 "name": "BaseBdev4", 00:34:01.113 "uuid": "f3e91a18-a294-5dc8-96a6-06657f9af8da", 00:34:01.113 "is_configured": true, 00:34:01.113 "data_offset": 0, 00:34:01.113 "data_size": 65536 00:34:01.113 } 00:34:01.113 ] 00:34:01.113 }' 00:34:01.113 13:54:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:01.113 13:54:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:34:01.113 13:54:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:01.113 13:54:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:34:01.113 13:54:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:34:02.485 13:54:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:34:02.485 13:54:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:02.485 13:54:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:02.485 13:54:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:02.485 13:54:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:02.485 13:54:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:02.485 13:54:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:02.485 13:54:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:02.485 13:54:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:34:02.485 13:54:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:02.485 13:54:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:02.485 13:54:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:02.485 "name": "raid_bdev1", 00:34:02.485 "uuid": "ff199959-08c2-4ac2-a4f9-969e7dd3ac3e", 00:34:02.485 "strip_size_kb": 64, 00:34:02.485 "state": "online", 00:34:02.485 "raid_level": "raid5f", 00:34:02.485 "superblock": false, 00:34:02.485 "num_base_bdevs": 4, 00:34:02.486 "num_base_bdevs_discovered": 4, 00:34:02.486 "num_base_bdevs_operational": 4, 00:34:02.486 "process": { 00:34:02.486 "type": "rebuild", 00:34:02.486 "target": "spare", 00:34:02.486 "progress": { 00:34:02.486 "blocks": 124800, 00:34:02.486 "percent": 63 00:34:02.486 } 00:34:02.486 }, 00:34:02.486 "base_bdevs_list": [ 00:34:02.486 { 00:34:02.486 "name": "spare", 00:34:02.486 "uuid": "78be2159-271b-5a17-bd0d-4657ec9e8724", 00:34:02.486 "is_configured": true, 00:34:02.486 "data_offset": 0, 00:34:02.486 "data_size": 65536 00:34:02.486 }, 00:34:02.486 { 00:34:02.486 "name": "BaseBdev2", 00:34:02.486 "uuid": "ef46d29f-fb56-5348-9e49-4a3825c02f93", 00:34:02.486 "is_configured": true, 00:34:02.486 "data_offset": 0, 00:34:02.486 "data_size": 65536 00:34:02.486 }, 00:34:02.486 { 00:34:02.486 "name": "BaseBdev3", 00:34:02.486 "uuid": "f2e1829c-1bf1-5ad2-845b-101076e9f258", 00:34:02.486 "is_configured": true, 00:34:02.486 "data_offset": 0, 00:34:02.486 "data_size": 65536 00:34:02.486 }, 00:34:02.486 { 00:34:02.486 "name": "BaseBdev4", 00:34:02.486 "uuid": "f3e91a18-a294-5dc8-96a6-06657f9af8da", 00:34:02.486 "is_configured": true, 00:34:02.486 "data_offset": 0, 00:34:02.486 "data_size": 65536 00:34:02.486 } 00:34:02.486 ] 00:34:02.486 }' 00:34:02.486 13:54:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:02.486 13:54:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:34:02.486 13:54:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:02.486 13:54:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:34:02.486 13:54:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:34:03.419 13:54:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:34:03.419 13:54:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:03.419 13:54:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:03.419 13:54:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:03.419 13:54:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:03.419 13:54:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:03.419 13:54:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:03.419 13:54:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:03.419 13:54:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:03.419 13:54:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:34:03.419 13:54:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:03.419 13:54:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:03.419 "name": "raid_bdev1", 00:34:03.419 "uuid": "ff199959-08c2-4ac2-a4f9-969e7dd3ac3e", 00:34:03.419 "strip_size_kb": 64, 00:34:03.419 "state": "online", 00:34:03.419 "raid_level": "raid5f", 00:34:03.419 "superblock": false, 00:34:03.419 "num_base_bdevs": 4, 00:34:03.419 "num_base_bdevs_discovered": 4, 00:34:03.419 "num_base_bdevs_operational": 4, 00:34:03.419 "process": { 00:34:03.419 "type": "rebuild", 00:34:03.419 "target": "spare", 00:34:03.419 "progress": { 00:34:03.419 "blocks": 145920, 00:34:03.419 "percent": 74 00:34:03.419 } 00:34:03.419 }, 00:34:03.419 "base_bdevs_list": [ 00:34:03.419 { 00:34:03.419 "name": "spare", 00:34:03.419 "uuid": "78be2159-271b-5a17-bd0d-4657ec9e8724", 00:34:03.419 "is_configured": true, 00:34:03.419 "data_offset": 0, 00:34:03.419 "data_size": 65536 00:34:03.419 }, 00:34:03.419 { 00:34:03.419 "name": "BaseBdev2", 00:34:03.419 "uuid": "ef46d29f-fb56-5348-9e49-4a3825c02f93", 00:34:03.419 "is_configured": true, 00:34:03.419 "data_offset": 0, 00:34:03.419 "data_size": 65536 00:34:03.419 }, 00:34:03.419 { 00:34:03.419 "name": "BaseBdev3", 00:34:03.419 "uuid": "f2e1829c-1bf1-5ad2-845b-101076e9f258", 00:34:03.420 "is_configured": true, 00:34:03.420 "data_offset": 0, 00:34:03.420 "data_size": 65536 00:34:03.420 }, 00:34:03.420 { 00:34:03.420 "name": "BaseBdev4", 00:34:03.420 "uuid": "f3e91a18-a294-5dc8-96a6-06657f9af8da", 00:34:03.420 "is_configured": true, 00:34:03.420 "data_offset": 0, 00:34:03.420 "data_size": 65536 00:34:03.420 } 00:34:03.420 ] 00:34:03.420 }' 00:34:03.420 13:54:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:03.420 13:54:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:34:03.420 13:54:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:03.420 13:54:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:34:03.420 13:54:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:34:04.353 13:54:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:34:04.353 13:54:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:04.353 13:54:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:04.353 13:54:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:04.353 13:54:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:04.353 13:54:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:04.353 13:54:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:04.353 13:54:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:04.353 13:54:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:04.353 13:54:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:34:04.353 13:54:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:04.353 13:54:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:04.353 "name": "raid_bdev1", 00:34:04.353 "uuid": "ff199959-08c2-4ac2-a4f9-969e7dd3ac3e", 00:34:04.353 "strip_size_kb": 64, 00:34:04.353 "state": "online", 00:34:04.353 "raid_level": "raid5f", 00:34:04.353 "superblock": false, 00:34:04.353 "num_base_bdevs": 4, 00:34:04.353 "num_base_bdevs_discovered": 4, 00:34:04.353 "num_base_bdevs_operational": 4, 00:34:04.353 "process": { 00:34:04.353 "type": "rebuild", 00:34:04.353 "target": "spare", 00:34:04.353 "progress": { 00:34:04.353 "blocks": 167040, 00:34:04.353 "percent": 84 00:34:04.353 } 00:34:04.353 }, 00:34:04.353 "base_bdevs_list": [ 00:34:04.353 { 00:34:04.353 "name": "spare", 00:34:04.353 "uuid": "78be2159-271b-5a17-bd0d-4657ec9e8724", 00:34:04.353 "is_configured": true, 00:34:04.353 "data_offset": 0, 00:34:04.353 "data_size": 65536 00:34:04.353 }, 00:34:04.353 { 00:34:04.353 "name": "BaseBdev2", 00:34:04.353 "uuid": "ef46d29f-fb56-5348-9e49-4a3825c02f93", 00:34:04.353 "is_configured": true, 00:34:04.353 "data_offset": 0, 00:34:04.353 "data_size": 65536 00:34:04.353 }, 00:34:04.353 { 00:34:04.353 "name": "BaseBdev3", 00:34:04.353 "uuid": "f2e1829c-1bf1-5ad2-845b-101076e9f258", 00:34:04.353 "is_configured": true, 00:34:04.353 "data_offset": 0, 00:34:04.353 "data_size": 65536 00:34:04.353 }, 00:34:04.353 { 00:34:04.353 "name": "BaseBdev4", 00:34:04.353 "uuid": "f3e91a18-a294-5dc8-96a6-06657f9af8da", 00:34:04.353 "is_configured": true, 00:34:04.353 "data_offset": 0, 00:34:04.353 "data_size": 65536 00:34:04.353 } 00:34:04.353 ] 00:34:04.353 }' 00:34:04.353 13:54:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:04.353 13:54:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:34:04.353 13:54:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:04.353 13:54:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:34:04.353 13:54:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:34:05.727 13:54:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:34:05.727 13:54:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:05.727 13:54:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:05.727 13:54:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:05.727 13:54:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:05.727 13:54:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:05.727 13:54:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:05.727 13:54:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:05.727 13:54:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:05.727 13:54:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:34:05.727 13:54:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:05.727 13:54:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:05.727 "name": "raid_bdev1", 00:34:05.727 "uuid": "ff199959-08c2-4ac2-a4f9-969e7dd3ac3e", 00:34:05.727 "strip_size_kb": 64, 00:34:05.727 "state": "online", 00:34:05.727 "raid_level": "raid5f", 00:34:05.727 "superblock": false, 00:34:05.727 "num_base_bdevs": 4, 00:34:05.727 "num_base_bdevs_discovered": 4, 00:34:05.727 "num_base_bdevs_operational": 4, 00:34:05.727 "process": { 00:34:05.727 "type": "rebuild", 00:34:05.727 "target": "spare", 00:34:05.727 "progress": { 00:34:05.727 "blocks": 188160, 00:34:05.727 "percent": 95 00:34:05.727 } 00:34:05.727 }, 00:34:05.727 "base_bdevs_list": [ 00:34:05.727 { 00:34:05.727 "name": "spare", 00:34:05.727 "uuid": "78be2159-271b-5a17-bd0d-4657ec9e8724", 00:34:05.727 "is_configured": true, 00:34:05.727 "data_offset": 0, 00:34:05.727 "data_size": 65536 00:34:05.727 }, 00:34:05.727 { 00:34:05.727 "name": "BaseBdev2", 00:34:05.727 "uuid": "ef46d29f-fb56-5348-9e49-4a3825c02f93", 00:34:05.727 "is_configured": true, 00:34:05.727 "data_offset": 0, 00:34:05.727 "data_size": 65536 00:34:05.727 }, 00:34:05.727 { 00:34:05.727 "name": "BaseBdev3", 00:34:05.727 "uuid": "f2e1829c-1bf1-5ad2-845b-101076e9f258", 00:34:05.727 "is_configured": true, 00:34:05.727 "data_offset": 0, 00:34:05.727 "data_size": 65536 00:34:05.727 }, 00:34:05.727 { 00:34:05.727 "name": "BaseBdev4", 00:34:05.727 "uuid": "f3e91a18-a294-5dc8-96a6-06657f9af8da", 00:34:05.727 "is_configured": true, 00:34:05.727 "data_offset": 0, 00:34:05.727 "data_size": 65536 00:34:05.727 } 00:34:05.727 ] 00:34:05.727 }' 00:34:05.727 13:54:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:05.727 13:54:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:34:05.727 13:54:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:05.727 13:54:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:34:05.727 13:54:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:34:05.991 [2024-11-20 13:54:12.845619] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:34:05.991 [2024-11-20 13:54:12.845712] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:34:05.991 [2024-11-20 13:54:12.845765] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:34:06.558 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:34:06.558 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:06.558 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:06.558 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:06.558 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:06.558 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:06.558 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:06.558 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:06.558 13:54:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:06.558 13:54:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:34:06.558 13:54:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:06.558 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:06.558 "name": "raid_bdev1", 00:34:06.558 "uuid": "ff199959-08c2-4ac2-a4f9-969e7dd3ac3e", 00:34:06.558 "strip_size_kb": 64, 00:34:06.558 "state": "online", 00:34:06.558 "raid_level": "raid5f", 00:34:06.558 "superblock": false, 00:34:06.558 "num_base_bdevs": 4, 00:34:06.558 "num_base_bdevs_discovered": 4, 00:34:06.558 "num_base_bdevs_operational": 4, 00:34:06.558 "base_bdevs_list": [ 00:34:06.558 { 00:34:06.558 "name": "spare", 00:34:06.558 "uuid": "78be2159-271b-5a17-bd0d-4657ec9e8724", 00:34:06.558 "is_configured": true, 00:34:06.558 "data_offset": 0, 00:34:06.558 "data_size": 65536 00:34:06.558 }, 00:34:06.558 { 00:34:06.558 "name": "BaseBdev2", 00:34:06.558 "uuid": "ef46d29f-fb56-5348-9e49-4a3825c02f93", 00:34:06.558 "is_configured": true, 00:34:06.558 "data_offset": 0, 00:34:06.558 "data_size": 65536 00:34:06.558 }, 00:34:06.558 { 00:34:06.558 "name": "BaseBdev3", 00:34:06.558 "uuid": "f2e1829c-1bf1-5ad2-845b-101076e9f258", 00:34:06.558 "is_configured": true, 00:34:06.558 "data_offset": 0, 00:34:06.558 "data_size": 65536 00:34:06.558 }, 00:34:06.558 { 00:34:06.558 "name": "BaseBdev4", 00:34:06.558 "uuid": "f3e91a18-a294-5dc8-96a6-06657f9af8da", 00:34:06.558 "is_configured": true, 00:34:06.558 "data_offset": 0, 00:34:06.558 "data_size": 65536 00:34:06.558 } 00:34:06.558 ] 00:34:06.558 }' 00:34:06.558 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:06.558 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:34:06.558 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:06.558 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:34:06.558 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:34:06.558 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:34:06.558 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:06.558 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:34:06.558 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:34:06.558 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:06.558 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:06.558 13:54:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:06.558 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:06.558 13:54:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:34:06.816 13:54:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:06.816 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:06.816 "name": "raid_bdev1", 00:34:06.816 "uuid": "ff199959-08c2-4ac2-a4f9-969e7dd3ac3e", 00:34:06.816 "strip_size_kb": 64, 00:34:06.816 "state": "online", 00:34:06.816 "raid_level": "raid5f", 00:34:06.816 "superblock": false, 00:34:06.816 "num_base_bdevs": 4, 00:34:06.816 "num_base_bdevs_discovered": 4, 00:34:06.816 "num_base_bdevs_operational": 4, 00:34:06.816 "base_bdevs_list": [ 00:34:06.816 { 00:34:06.816 "name": "spare", 00:34:06.816 "uuid": "78be2159-271b-5a17-bd0d-4657ec9e8724", 00:34:06.816 "is_configured": true, 00:34:06.816 "data_offset": 0, 00:34:06.816 "data_size": 65536 00:34:06.816 }, 00:34:06.816 { 00:34:06.816 "name": "BaseBdev2", 00:34:06.816 "uuid": "ef46d29f-fb56-5348-9e49-4a3825c02f93", 00:34:06.816 "is_configured": true, 00:34:06.816 "data_offset": 0, 00:34:06.816 "data_size": 65536 00:34:06.816 }, 00:34:06.816 { 00:34:06.816 "name": "BaseBdev3", 00:34:06.816 "uuid": "f2e1829c-1bf1-5ad2-845b-101076e9f258", 00:34:06.816 "is_configured": true, 00:34:06.816 "data_offset": 0, 00:34:06.816 "data_size": 65536 00:34:06.816 }, 00:34:06.816 { 00:34:06.816 "name": "BaseBdev4", 00:34:06.816 "uuid": "f3e91a18-a294-5dc8-96a6-06657f9af8da", 00:34:06.816 "is_configured": true, 00:34:06.816 "data_offset": 0, 00:34:06.816 "data_size": 65536 00:34:06.816 } 00:34:06.816 ] 00:34:06.816 }' 00:34:06.816 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:06.816 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:34:06.816 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:06.816 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:34:06.816 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:34:06.816 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:06.816 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:06.816 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:34:06.816 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:34:06.816 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:34:06.816 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:06.816 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:06.817 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:06.817 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:06.817 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:06.817 13:54:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:06.817 13:54:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:34:06.817 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:06.817 13:54:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:06.817 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:06.817 "name": "raid_bdev1", 00:34:06.817 "uuid": "ff199959-08c2-4ac2-a4f9-969e7dd3ac3e", 00:34:06.817 "strip_size_kb": 64, 00:34:06.817 "state": "online", 00:34:06.817 "raid_level": "raid5f", 00:34:06.817 "superblock": false, 00:34:06.817 "num_base_bdevs": 4, 00:34:06.817 "num_base_bdevs_discovered": 4, 00:34:06.817 "num_base_bdevs_operational": 4, 00:34:06.817 "base_bdevs_list": [ 00:34:06.817 { 00:34:06.817 "name": "spare", 00:34:06.817 "uuid": "78be2159-271b-5a17-bd0d-4657ec9e8724", 00:34:06.817 "is_configured": true, 00:34:06.817 "data_offset": 0, 00:34:06.817 "data_size": 65536 00:34:06.817 }, 00:34:06.817 { 00:34:06.817 "name": "BaseBdev2", 00:34:06.817 "uuid": "ef46d29f-fb56-5348-9e49-4a3825c02f93", 00:34:06.817 "is_configured": true, 00:34:06.817 "data_offset": 0, 00:34:06.817 "data_size": 65536 00:34:06.817 }, 00:34:06.817 { 00:34:06.817 "name": "BaseBdev3", 00:34:06.817 "uuid": "f2e1829c-1bf1-5ad2-845b-101076e9f258", 00:34:06.817 "is_configured": true, 00:34:06.817 "data_offset": 0, 00:34:06.817 "data_size": 65536 00:34:06.817 }, 00:34:06.817 { 00:34:06.817 "name": "BaseBdev4", 00:34:06.817 "uuid": "f3e91a18-a294-5dc8-96a6-06657f9af8da", 00:34:06.817 "is_configured": true, 00:34:06.817 "data_offset": 0, 00:34:06.817 "data_size": 65536 00:34:06.817 } 00:34:06.817 ] 00:34:06.817 }' 00:34:06.817 13:54:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:06.817 13:54:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:34:07.076 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:34:07.076 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:07.076 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:34:07.076 [2024-11-20 13:54:14.023203] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:34:07.076 [2024-11-20 13:54:14.023377] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:34:07.076 [2024-11-20 13:54:14.023474] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:34:07.076 [2024-11-20 13:54:14.023570] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:34:07.076 [2024-11-20 13:54:14.023580] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:34:07.076 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:07.076 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:07.076 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:07.076 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:34:07.076 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:34:07.076 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:07.076 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:34:07.076 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:34:07.076 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:34:07.076 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:34:07.076 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:34:07.076 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:34:07.076 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:34:07.076 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:34:07.076 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:34:07.076 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:34:07.076 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:34:07.076 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:34:07.076 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:34:07.333 /dev/nbd0 00:34:07.333 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:34:07.333 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:34:07.333 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:34:07.333 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:34:07.333 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:34:07.333 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:34:07.333 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:34:07.333 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:34:07.333 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:34:07.333 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:34:07.333 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:34:07.333 1+0 records in 00:34:07.333 1+0 records out 00:34:07.333 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000186752 s, 21.9 MB/s 00:34:07.333 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:34:07.333 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:34:07.333 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:34:07.333 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:34:07.333 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:34:07.333 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:34:07.333 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:34:07.333 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:34:07.591 /dev/nbd1 00:34:07.591 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:34:07.591 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:34:07.591 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:34:07.591 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:34:07.591 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:34:07.591 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:34:07.591 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:34:07.591 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:34:07.591 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:34:07.591 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:34:07.591 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:34:07.591 1+0 records in 00:34:07.591 1+0 records out 00:34:07.591 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000262022 s, 15.6 MB/s 00:34:07.591 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:34:07.591 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:34:07.591 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:34:07.591 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:34:07.591 13:54:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:34:07.591 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:34:07.591 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:34:07.591 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:34:07.591 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:34:07.591 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:34:07.591 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:34:07.591 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:34:07.591 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:34:07.591 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:34:07.591 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:34:07.849 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:34:07.849 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:34:07.849 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:34:07.849 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:34:07.849 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:34:07.849 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:34:07.849 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:34:07.849 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:34:07.849 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:34:07.849 13:54:14 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:34:08.108 13:54:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:34:08.108 13:54:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:34:08.108 13:54:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:34:08.108 13:54:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:34:08.108 13:54:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:34:08.108 13:54:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:34:08.108 13:54:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:34:08.108 13:54:15 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:34:08.108 13:54:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:34:08.108 13:54:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 82119 00:34:08.108 13:54:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 82119 ']' 00:34:08.108 13:54:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 82119 00:34:08.108 13:54:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:34:08.108 13:54:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:34:08.108 13:54:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 82119 00:34:08.108 13:54:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:34:08.108 13:54:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:34:08.108 killing process with pid 82119 00:34:08.108 13:54:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 82119' 00:34:08.108 13:54:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@973 -- # kill 82119 00:34:08.108 Received shutdown signal, test time was about 60.000000 seconds 00:34:08.108 00:34:08.108 Latency(us) 00:34:08.108 [2024-11-20T13:54:15.167Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:34:08.108 [2024-11-20T13:54:15.167Z] =================================================================================================================== 00:34:08.108 [2024-11-20T13:54:15.167Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:34:08.108 13:54:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@978 -- # wait 82119 00:34:08.108 [2024-11-20 13:54:15.134929] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:34:08.367 [2024-11-20 13:54:15.388388] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:34:09.301 13:54:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:34:09.301 00:34:09.301 real 0m17.993s 00:34:09.301 user 0m20.974s 00:34:09.301 sys 0m1.848s 00:34:09.301 13:54:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:34:09.301 13:54:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:34:09.301 ************************************ 00:34:09.301 END TEST raid5f_rebuild_test 00:34:09.301 ************************************ 00:34:09.301 13:54:16 bdev_raid -- bdev/bdev_raid.sh@991 -- # run_test raid5f_rebuild_test_sb raid_rebuild_test raid5f 4 true false true 00:34:09.301 13:54:16 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:34:09.301 13:54:16 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:34:09.301 13:54:16 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:34:09.301 ************************************ 00:34:09.301 START TEST raid5f_rebuild_test_sb 00:34:09.301 ************************************ 00:34:09.301 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 4 true false true 00:34:09.301 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:34:09.301 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:34:09.301 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:34:09.301 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:34:09.301 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:34:09.301 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:34:09.301 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:34:09.301 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:34:09.301 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:34:09.301 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:34:09.301 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:34:09.301 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:34:09.301 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:34:09.301 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=82619 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 82619 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 82619 ']' 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:34:09.302 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:09.302 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:34:09.302 [2024-11-20 13:54:16.114241] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:34:09.302 I/O size of 3145728 is greater than zero copy threshold (65536). 00:34:09.302 Zero copy mechanism will not be used. 00:34:09.302 [2024-11-20 13:54:16.114370] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid82619 ] 00:34:09.302 [2024-11-20 13:54:16.281031] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:34:09.560 [2024-11-20 13:54:16.396793] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:34:09.560 [2024-11-20 13:54:16.544489] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:34:09.560 [2024-11-20 13:54:16.544532] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:34:10.127 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:34:10.127 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:34:10.127 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:34:10.127 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:34:10.127 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:10.127 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:10.127 BaseBdev1_malloc 00:34:10.127 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:10.127 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:34:10.127 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:10.127 13:54:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:10.127 [2024-11-20 13:54:17.002282] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:34:10.127 [2024-11-20 13:54:17.002350] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:34:10.127 [2024-11-20 13:54:17.002374] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:34:10.127 [2024-11-20 13:54:17.002386] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:34:10.127 [2024-11-20 13:54:17.004675] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:34:10.127 [2024-11-20 13:54:17.004713] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:34:10.127 BaseBdev1 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:10.128 BaseBdev2_malloc 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:10.128 [2024-11-20 13:54:17.040061] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:34:10.128 [2024-11-20 13:54:17.040115] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:34:10.128 [2024-11-20 13:54:17.040136] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:34:10.128 [2024-11-20 13:54:17.040148] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:34:10.128 [2024-11-20 13:54:17.042416] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:34:10.128 [2024-11-20 13:54:17.042451] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:34:10.128 BaseBdev2 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:10.128 BaseBdev3_malloc 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:10.128 [2024-11-20 13:54:17.085839] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:34:10.128 [2024-11-20 13:54:17.085895] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:34:10.128 [2024-11-20 13:54:17.085917] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:34:10.128 [2024-11-20 13:54:17.085929] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:34:10.128 [2024-11-20 13:54:17.088203] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:34:10.128 [2024-11-20 13:54:17.088240] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:34:10.128 BaseBdev3 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:10.128 BaseBdev4_malloc 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:10.128 [2024-11-20 13:54:17.123966] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:34:10.128 [2024-11-20 13:54:17.124028] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:34:10.128 [2024-11-20 13:54:17.124046] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:34:10.128 [2024-11-20 13:54:17.124057] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:34:10.128 [2024-11-20 13:54:17.126234] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:34:10.128 [2024-11-20 13:54:17.126271] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:34:10.128 BaseBdev4 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:10.128 spare_malloc 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:10.128 spare_delay 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:10.128 [2024-11-20 13:54:17.169807] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:34:10.128 [2024-11-20 13:54:17.169856] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:34:10.128 [2024-11-20 13:54:17.169874] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:34:10.128 [2024-11-20 13:54:17.169885] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:34:10.128 [2024-11-20 13:54:17.172066] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:34:10.128 [2024-11-20 13:54:17.172099] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:34:10.128 spare 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:10.128 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:10.128 [2024-11-20 13:54:17.177878] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:34:10.128 [2024-11-20 13:54:17.179806] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:34:10.128 [2024-11-20 13:54:17.179870] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:34:10.128 [2024-11-20 13:54:17.179922] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:34:10.128 [2024-11-20 13:54:17.180115] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:34:10.128 [2024-11-20 13:54:17.180137] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:34:10.128 [2024-11-20 13:54:17.180385] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:34:10.387 [2024-11-20 13:54:17.185381] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:34:10.387 [2024-11-20 13:54:17.185401] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:34:10.387 [2024-11-20 13:54:17.185571] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:34:10.387 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:10.387 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:34:10.387 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:10.387 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:10.387 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:34:10.387 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:34:10.387 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:34:10.387 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:10.387 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:10.387 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:10.387 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:10.387 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:10.387 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:10.387 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:10.387 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:10.387 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:10.387 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:10.387 "name": "raid_bdev1", 00:34:10.387 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:10.387 "strip_size_kb": 64, 00:34:10.387 "state": "online", 00:34:10.387 "raid_level": "raid5f", 00:34:10.387 "superblock": true, 00:34:10.387 "num_base_bdevs": 4, 00:34:10.387 "num_base_bdevs_discovered": 4, 00:34:10.387 "num_base_bdevs_operational": 4, 00:34:10.387 "base_bdevs_list": [ 00:34:10.387 { 00:34:10.387 "name": "BaseBdev1", 00:34:10.387 "uuid": "4aacc9c3-99a5-574a-9a88-e56cc41aa6d4", 00:34:10.387 "is_configured": true, 00:34:10.387 "data_offset": 2048, 00:34:10.387 "data_size": 63488 00:34:10.387 }, 00:34:10.387 { 00:34:10.387 "name": "BaseBdev2", 00:34:10.387 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:10.387 "is_configured": true, 00:34:10.387 "data_offset": 2048, 00:34:10.387 "data_size": 63488 00:34:10.387 }, 00:34:10.387 { 00:34:10.387 "name": "BaseBdev3", 00:34:10.387 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:10.387 "is_configured": true, 00:34:10.387 "data_offset": 2048, 00:34:10.387 "data_size": 63488 00:34:10.387 }, 00:34:10.387 { 00:34:10.387 "name": "BaseBdev4", 00:34:10.387 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:10.387 "is_configured": true, 00:34:10.387 "data_offset": 2048, 00:34:10.387 "data_size": 63488 00:34:10.387 } 00:34:10.387 ] 00:34:10.387 }' 00:34:10.387 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:10.387 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:10.645 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:34:10.645 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:10.645 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:10.645 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:34:10.645 [2024-11-20 13:54:17.507530] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:34:10.645 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:10.645 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=190464 00:34:10.645 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:10.645 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:34:10.645 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:10.645 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:10.645 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:10.645 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:34:10.645 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:34:10.645 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:34:10.645 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:34:10.645 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:34:10.645 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:34:10.645 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:34:10.645 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:34:10.645 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:34:10.645 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:34:10.645 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:34:10.645 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:34:10.645 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:34:10.645 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:34:10.904 [2024-11-20 13:54:17.711380] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:34:10.904 /dev/nbd0 00:34:10.904 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:34:10.904 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:34:10.904 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:34:10.904 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:34:10.904 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:34:10.904 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:34:10.904 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:34:10.904 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:34:10.904 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:34:10.904 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:34:10.904 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:34:10.904 1+0 records in 00:34:10.904 1+0 records out 00:34:10.904 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000130418 s, 31.4 MB/s 00:34:10.904 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:34:10.904 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:34:10.904 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:34:10.904 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:34:10.904 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:34:10.904 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:34:10.904 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:34:10.904 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:34:10.904 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@630 -- # write_unit_size=384 00:34:10.904 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@631 -- # echo 192 00:34:10.904 13:54:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=196608 count=496 oflag=direct 00:34:11.161 496+0 records in 00:34:11.161 496+0 records out 00:34:11.161 97517568 bytes (98 MB, 93 MiB) copied, 0.41326 s, 236 MB/s 00:34:11.161 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:34:11.161 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:34:11.161 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:34:11.161 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:34:11.161 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:34:11.161 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:34:11.161 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:34:11.420 [2024-11-20 13:54:18.345625] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:11.420 [2024-11-20 13:54:18.355319] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:11.420 "name": "raid_bdev1", 00:34:11.420 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:11.420 "strip_size_kb": 64, 00:34:11.420 "state": "online", 00:34:11.420 "raid_level": "raid5f", 00:34:11.420 "superblock": true, 00:34:11.420 "num_base_bdevs": 4, 00:34:11.420 "num_base_bdevs_discovered": 3, 00:34:11.420 "num_base_bdevs_operational": 3, 00:34:11.420 "base_bdevs_list": [ 00:34:11.420 { 00:34:11.420 "name": null, 00:34:11.420 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:11.420 "is_configured": false, 00:34:11.420 "data_offset": 0, 00:34:11.420 "data_size": 63488 00:34:11.420 }, 00:34:11.420 { 00:34:11.420 "name": "BaseBdev2", 00:34:11.420 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:11.420 "is_configured": true, 00:34:11.420 "data_offset": 2048, 00:34:11.420 "data_size": 63488 00:34:11.420 }, 00:34:11.420 { 00:34:11.420 "name": "BaseBdev3", 00:34:11.420 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:11.420 "is_configured": true, 00:34:11.420 "data_offset": 2048, 00:34:11.420 "data_size": 63488 00:34:11.420 }, 00:34:11.420 { 00:34:11.420 "name": "BaseBdev4", 00:34:11.420 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:11.420 "is_configured": true, 00:34:11.420 "data_offset": 2048, 00:34:11.420 "data_size": 63488 00:34:11.420 } 00:34:11.420 ] 00:34:11.420 }' 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:11.420 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:11.679 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:34:11.679 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:11.679 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:11.679 [2024-11-20 13:54:18.647387] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:34:11.679 [2024-11-20 13:54:18.657758] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002aa50 00:34:11.679 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:11.679 13:54:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:34:11.679 [2024-11-20 13:54:18.664680] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:34:12.722 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:12.722 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:12.722 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:12.722 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:12.722 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:12.722 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:12.722 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:12.722 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:12.722 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:12.722 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:12.722 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:12.722 "name": "raid_bdev1", 00:34:12.722 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:12.722 "strip_size_kb": 64, 00:34:12.722 "state": "online", 00:34:12.722 "raid_level": "raid5f", 00:34:12.722 "superblock": true, 00:34:12.722 "num_base_bdevs": 4, 00:34:12.722 "num_base_bdevs_discovered": 4, 00:34:12.722 "num_base_bdevs_operational": 4, 00:34:12.722 "process": { 00:34:12.722 "type": "rebuild", 00:34:12.722 "target": "spare", 00:34:12.722 "progress": { 00:34:12.722 "blocks": 17280, 00:34:12.722 "percent": 9 00:34:12.722 } 00:34:12.722 }, 00:34:12.722 "base_bdevs_list": [ 00:34:12.722 { 00:34:12.722 "name": "spare", 00:34:12.722 "uuid": "cdf0ebd9-45ac-5c95-8cc7-c198588e1fef", 00:34:12.722 "is_configured": true, 00:34:12.722 "data_offset": 2048, 00:34:12.722 "data_size": 63488 00:34:12.722 }, 00:34:12.722 { 00:34:12.722 "name": "BaseBdev2", 00:34:12.722 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:12.722 "is_configured": true, 00:34:12.722 "data_offset": 2048, 00:34:12.722 "data_size": 63488 00:34:12.722 }, 00:34:12.722 { 00:34:12.722 "name": "BaseBdev3", 00:34:12.722 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:12.722 "is_configured": true, 00:34:12.722 "data_offset": 2048, 00:34:12.722 "data_size": 63488 00:34:12.722 }, 00:34:12.722 { 00:34:12.722 "name": "BaseBdev4", 00:34:12.722 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:12.722 "is_configured": true, 00:34:12.722 "data_offset": 2048, 00:34:12.722 "data_size": 63488 00:34:12.722 } 00:34:12.722 ] 00:34:12.722 }' 00:34:12.722 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:12.722 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:34:12.722 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:12.722 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:34:12.722 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:34:12.722 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:12.722 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:12.722 [2024-11-20 13:54:19.770139] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:34:12.722 [2024-11-20 13:54:19.774283] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:34:12.722 [2024-11-20 13:54:19.774355] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:34:12.722 [2024-11-20 13:54:19.774373] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:34:12.722 [2024-11-20 13:54:19.774384] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:34:12.981 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:12.982 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:34:12.982 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:12.982 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:12.982 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:34:12.982 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:34:12.982 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:34:12.982 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:12.982 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:12.982 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:12.982 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:12.982 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:12.982 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:12.982 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:12.982 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:12.982 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:12.982 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:12.982 "name": "raid_bdev1", 00:34:12.982 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:12.982 "strip_size_kb": 64, 00:34:12.982 "state": "online", 00:34:12.982 "raid_level": "raid5f", 00:34:12.982 "superblock": true, 00:34:12.982 "num_base_bdevs": 4, 00:34:12.982 "num_base_bdevs_discovered": 3, 00:34:12.982 "num_base_bdevs_operational": 3, 00:34:12.982 "base_bdevs_list": [ 00:34:12.982 { 00:34:12.982 "name": null, 00:34:12.982 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:12.982 "is_configured": false, 00:34:12.982 "data_offset": 0, 00:34:12.982 "data_size": 63488 00:34:12.982 }, 00:34:12.982 { 00:34:12.982 "name": "BaseBdev2", 00:34:12.982 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:12.982 "is_configured": true, 00:34:12.982 "data_offset": 2048, 00:34:12.982 "data_size": 63488 00:34:12.982 }, 00:34:12.982 { 00:34:12.982 "name": "BaseBdev3", 00:34:12.982 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:12.982 "is_configured": true, 00:34:12.982 "data_offset": 2048, 00:34:12.982 "data_size": 63488 00:34:12.982 }, 00:34:12.982 { 00:34:12.982 "name": "BaseBdev4", 00:34:12.982 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:12.982 "is_configured": true, 00:34:12.982 "data_offset": 2048, 00:34:12.982 "data_size": 63488 00:34:12.982 } 00:34:12.982 ] 00:34:12.982 }' 00:34:12.982 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:12.982 13:54:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:13.241 13:54:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:34:13.241 13:54:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:13.241 13:54:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:34:13.241 13:54:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:34:13.241 13:54:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:13.241 13:54:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:13.241 13:54:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:13.241 13:54:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:13.241 13:54:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:13.241 13:54:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:13.241 13:54:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:13.241 "name": "raid_bdev1", 00:34:13.241 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:13.241 "strip_size_kb": 64, 00:34:13.241 "state": "online", 00:34:13.241 "raid_level": "raid5f", 00:34:13.241 "superblock": true, 00:34:13.241 "num_base_bdevs": 4, 00:34:13.241 "num_base_bdevs_discovered": 3, 00:34:13.241 "num_base_bdevs_operational": 3, 00:34:13.241 "base_bdevs_list": [ 00:34:13.241 { 00:34:13.241 "name": null, 00:34:13.241 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:13.241 "is_configured": false, 00:34:13.241 "data_offset": 0, 00:34:13.241 "data_size": 63488 00:34:13.241 }, 00:34:13.241 { 00:34:13.241 "name": "BaseBdev2", 00:34:13.241 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:13.241 "is_configured": true, 00:34:13.241 "data_offset": 2048, 00:34:13.241 "data_size": 63488 00:34:13.241 }, 00:34:13.241 { 00:34:13.241 "name": "BaseBdev3", 00:34:13.241 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:13.241 "is_configured": true, 00:34:13.241 "data_offset": 2048, 00:34:13.241 "data_size": 63488 00:34:13.241 }, 00:34:13.241 { 00:34:13.241 "name": "BaseBdev4", 00:34:13.241 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:13.241 "is_configured": true, 00:34:13.241 "data_offset": 2048, 00:34:13.241 "data_size": 63488 00:34:13.241 } 00:34:13.241 ] 00:34:13.241 }' 00:34:13.241 13:54:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:13.241 13:54:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:34:13.241 13:54:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:13.241 13:54:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:34:13.241 13:54:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:34:13.241 13:54:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:13.241 13:54:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:13.241 [2024-11-20 13:54:20.214801] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:34:13.241 [2024-11-20 13:54:20.224742] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002ab20 00:34:13.241 13:54:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:13.241 13:54:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:34:13.241 [2024-11-20 13:54:20.231376] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:34:14.175 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:14.175 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:14.175 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:14.175 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:14.175 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:14.433 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:14.433 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:14.433 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:14.433 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:14.433 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:14.433 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:14.433 "name": "raid_bdev1", 00:34:14.433 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:14.433 "strip_size_kb": 64, 00:34:14.433 "state": "online", 00:34:14.433 "raid_level": "raid5f", 00:34:14.433 "superblock": true, 00:34:14.433 "num_base_bdevs": 4, 00:34:14.433 "num_base_bdevs_discovered": 4, 00:34:14.433 "num_base_bdevs_operational": 4, 00:34:14.433 "process": { 00:34:14.433 "type": "rebuild", 00:34:14.433 "target": "spare", 00:34:14.434 "progress": { 00:34:14.434 "blocks": 19200, 00:34:14.434 "percent": 10 00:34:14.434 } 00:34:14.434 }, 00:34:14.434 "base_bdevs_list": [ 00:34:14.434 { 00:34:14.434 "name": "spare", 00:34:14.434 "uuid": "cdf0ebd9-45ac-5c95-8cc7-c198588e1fef", 00:34:14.434 "is_configured": true, 00:34:14.434 "data_offset": 2048, 00:34:14.434 "data_size": 63488 00:34:14.434 }, 00:34:14.434 { 00:34:14.434 "name": "BaseBdev2", 00:34:14.434 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:14.434 "is_configured": true, 00:34:14.434 "data_offset": 2048, 00:34:14.434 "data_size": 63488 00:34:14.434 }, 00:34:14.434 { 00:34:14.434 "name": "BaseBdev3", 00:34:14.434 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:14.434 "is_configured": true, 00:34:14.434 "data_offset": 2048, 00:34:14.434 "data_size": 63488 00:34:14.434 }, 00:34:14.434 { 00:34:14.434 "name": "BaseBdev4", 00:34:14.434 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:14.434 "is_configured": true, 00:34:14.434 "data_offset": 2048, 00:34:14.434 "data_size": 63488 00:34:14.434 } 00:34:14.434 ] 00:34:14.434 }' 00:34:14.434 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:14.434 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:34:14.434 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:14.434 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:34:14.434 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:34:14.434 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:34:14.434 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:34:14.434 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:34:14.434 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:34:14.434 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=508 00:34:14.434 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:34:14.434 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:14.434 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:14.434 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:14.434 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:14.434 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:14.434 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:14.434 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:14.434 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:14.434 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:14.434 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:14.434 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:14.434 "name": "raid_bdev1", 00:34:14.434 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:14.434 "strip_size_kb": 64, 00:34:14.434 "state": "online", 00:34:14.434 "raid_level": "raid5f", 00:34:14.434 "superblock": true, 00:34:14.434 "num_base_bdevs": 4, 00:34:14.434 "num_base_bdevs_discovered": 4, 00:34:14.434 "num_base_bdevs_operational": 4, 00:34:14.434 "process": { 00:34:14.434 "type": "rebuild", 00:34:14.434 "target": "spare", 00:34:14.434 "progress": { 00:34:14.434 "blocks": 21120, 00:34:14.434 "percent": 11 00:34:14.434 } 00:34:14.434 }, 00:34:14.434 "base_bdevs_list": [ 00:34:14.434 { 00:34:14.434 "name": "spare", 00:34:14.434 "uuid": "cdf0ebd9-45ac-5c95-8cc7-c198588e1fef", 00:34:14.434 "is_configured": true, 00:34:14.434 "data_offset": 2048, 00:34:14.434 "data_size": 63488 00:34:14.434 }, 00:34:14.434 { 00:34:14.434 "name": "BaseBdev2", 00:34:14.434 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:14.434 "is_configured": true, 00:34:14.434 "data_offset": 2048, 00:34:14.434 "data_size": 63488 00:34:14.434 }, 00:34:14.434 { 00:34:14.434 "name": "BaseBdev3", 00:34:14.434 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:14.434 "is_configured": true, 00:34:14.434 "data_offset": 2048, 00:34:14.434 "data_size": 63488 00:34:14.434 }, 00:34:14.434 { 00:34:14.434 "name": "BaseBdev4", 00:34:14.434 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:14.434 "is_configured": true, 00:34:14.434 "data_offset": 2048, 00:34:14.434 "data_size": 63488 00:34:14.434 } 00:34:14.434 ] 00:34:14.434 }' 00:34:14.434 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:14.434 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:34:14.434 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:14.434 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:34:14.434 13:54:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:34:15.808 13:54:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:34:15.808 13:54:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:15.808 13:54:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:15.808 13:54:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:15.808 13:54:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:15.808 13:54:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:15.808 13:54:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:15.808 13:54:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:15.808 13:54:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:15.808 13:54:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:15.808 13:54:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:15.808 13:54:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:15.808 "name": "raid_bdev1", 00:34:15.808 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:15.808 "strip_size_kb": 64, 00:34:15.808 "state": "online", 00:34:15.808 "raid_level": "raid5f", 00:34:15.808 "superblock": true, 00:34:15.808 "num_base_bdevs": 4, 00:34:15.808 "num_base_bdevs_discovered": 4, 00:34:15.808 "num_base_bdevs_operational": 4, 00:34:15.808 "process": { 00:34:15.808 "type": "rebuild", 00:34:15.808 "target": "spare", 00:34:15.808 "progress": { 00:34:15.808 "blocks": 40320, 00:34:15.808 "percent": 21 00:34:15.808 } 00:34:15.808 }, 00:34:15.808 "base_bdevs_list": [ 00:34:15.808 { 00:34:15.808 "name": "spare", 00:34:15.808 "uuid": "cdf0ebd9-45ac-5c95-8cc7-c198588e1fef", 00:34:15.808 "is_configured": true, 00:34:15.808 "data_offset": 2048, 00:34:15.808 "data_size": 63488 00:34:15.808 }, 00:34:15.808 { 00:34:15.808 "name": "BaseBdev2", 00:34:15.808 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:15.808 "is_configured": true, 00:34:15.808 "data_offset": 2048, 00:34:15.808 "data_size": 63488 00:34:15.808 }, 00:34:15.808 { 00:34:15.808 "name": "BaseBdev3", 00:34:15.808 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:15.809 "is_configured": true, 00:34:15.809 "data_offset": 2048, 00:34:15.809 "data_size": 63488 00:34:15.809 }, 00:34:15.809 { 00:34:15.809 "name": "BaseBdev4", 00:34:15.809 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:15.809 "is_configured": true, 00:34:15.809 "data_offset": 2048, 00:34:15.809 "data_size": 63488 00:34:15.809 } 00:34:15.809 ] 00:34:15.809 }' 00:34:15.809 13:54:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:15.809 13:54:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:34:15.809 13:54:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:15.809 13:54:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:34:15.809 13:54:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:34:16.743 13:54:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:34:16.743 13:54:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:16.743 13:54:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:16.743 13:54:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:16.743 13:54:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:16.743 13:54:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:16.743 13:54:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:16.743 13:54:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:16.743 13:54:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:16.743 13:54:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:16.743 13:54:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:16.743 13:54:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:16.743 "name": "raid_bdev1", 00:34:16.744 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:16.744 "strip_size_kb": 64, 00:34:16.744 "state": "online", 00:34:16.744 "raid_level": "raid5f", 00:34:16.744 "superblock": true, 00:34:16.744 "num_base_bdevs": 4, 00:34:16.744 "num_base_bdevs_discovered": 4, 00:34:16.744 "num_base_bdevs_operational": 4, 00:34:16.744 "process": { 00:34:16.744 "type": "rebuild", 00:34:16.744 "target": "spare", 00:34:16.744 "progress": { 00:34:16.744 "blocks": 61440, 00:34:16.744 "percent": 32 00:34:16.744 } 00:34:16.744 }, 00:34:16.744 "base_bdevs_list": [ 00:34:16.744 { 00:34:16.744 "name": "spare", 00:34:16.744 "uuid": "cdf0ebd9-45ac-5c95-8cc7-c198588e1fef", 00:34:16.744 "is_configured": true, 00:34:16.744 "data_offset": 2048, 00:34:16.744 "data_size": 63488 00:34:16.744 }, 00:34:16.744 { 00:34:16.744 "name": "BaseBdev2", 00:34:16.744 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:16.744 "is_configured": true, 00:34:16.744 "data_offset": 2048, 00:34:16.744 "data_size": 63488 00:34:16.744 }, 00:34:16.744 { 00:34:16.744 "name": "BaseBdev3", 00:34:16.744 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:16.744 "is_configured": true, 00:34:16.744 "data_offset": 2048, 00:34:16.744 "data_size": 63488 00:34:16.744 }, 00:34:16.744 { 00:34:16.744 "name": "BaseBdev4", 00:34:16.744 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:16.744 "is_configured": true, 00:34:16.744 "data_offset": 2048, 00:34:16.744 "data_size": 63488 00:34:16.744 } 00:34:16.744 ] 00:34:16.744 }' 00:34:16.744 13:54:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:16.744 13:54:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:34:16.744 13:54:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:16.744 13:54:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:34:16.744 13:54:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:34:17.675 13:54:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:34:17.675 13:54:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:17.675 13:54:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:17.675 13:54:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:17.675 13:54:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:17.675 13:54:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:17.675 13:54:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:17.675 13:54:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:17.675 13:54:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:17.675 13:54:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:17.676 13:54:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:17.676 13:54:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:17.676 "name": "raid_bdev1", 00:34:17.676 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:17.676 "strip_size_kb": 64, 00:34:17.676 "state": "online", 00:34:17.676 "raid_level": "raid5f", 00:34:17.676 "superblock": true, 00:34:17.676 "num_base_bdevs": 4, 00:34:17.676 "num_base_bdevs_discovered": 4, 00:34:17.676 "num_base_bdevs_operational": 4, 00:34:17.676 "process": { 00:34:17.676 "type": "rebuild", 00:34:17.676 "target": "spare", 00:34:17.676 "progress": { 00:34:17.676 "blocks": 82560, 00:34:17.676 "percent": 43 00:34:17.676 } 00:34:17.676 }, 00:34:17.676 "base_bdevs_list": [ 00:34:17.676 { 00:34:17.676 "name": "spare", 00:34:17.676 "uuid": "cdf0ebd9-45ac-5c95-8cc7-c198588e1fef", 00:34:17.676 "is_configured": true, 00:34:17.676 "data_offset": 2048, 00:34:17.676 "data_size": 63488 00:34:17.676 }, 00:34:17.676 { 00:34:17.676 "name": "BaseBdev2", 00:34:17.676 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:17.676 "is_configured": true, 00:34:17.676 "data_offset": 2048, 00:34:17.676 "data_size": 63488 00:34:17.676 }, 00:34:17.676 { 00:34:17.676 "name": "BaseBdev3", 00:34:17.676 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:17.676 "is_configured": true, 00:34:17.676 "data_offset": 2048, 00:34:17.676 "data_size": 63488 00:34:17.676 }, 00:34:17.676 { 00:34:17.676 "name": "BaseBdev4", 00:34:17.676 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:17.676 "is_configured": true, 00:34:17.676 "data_offset": 2048, 00:34:17.676 "data_size": 63488 00:34:17.676 } 00:34:17.676 ] 00:34:17.676 }' 00:34:17.676 13:54:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:17.676 13:54:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:34:17.676 13:54:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:17.676 13:54:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:34:17.676 13:54:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:34:19.113 13:54:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:34:19.113 13:54:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:19.113 13:54:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:19.113 13:54:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:19.113 13:54:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:19.113 13:54:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:19.113 13:54:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:19.113 13:54:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:19.113 13:54:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:19.113 13:54:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:19.113 13:54:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:19.113 13:54:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:19.113 "name": "raid_bdev1", 00:34:19.113 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:19.113 "strip_size_kb": 64, 00:34:19.113 "state": "online", 00:34:19.113 "raid_level": "raid5f", 00:34:19.113 "superblock": true, 00:34:19.113 "num_base_bdevs": 4, 00:34:19.113 "num_base_bdevs_discovered": 4, 00:34:19.113 "num_base_bdevs_operational": 4, 00:34:19.113 "process": { 00:34:19.113 "type": "rebuild", 00:34:19.113 "target": "spare", 00:34:19.113 "progress": { 00:34:19.113 "blocks": 103680, 00:34:19.113 "percent": 54 00:34:19.113 } 00:34:19.113 }, 00:34:19.113 "base_bdevs_list": [ 00:34:19.113 { 00:34:19.113 "name": "spare", 00:34:19.113 "uuid": "cdf0ebd9-45ac-5c95-8cc7-c198588e1fef", 00:34:19.113 "is_configured": true, 00:34:19.113 "data_offset": 2048, 00:34:19.113 "data_size": 63488 00:34:19.113 }, 00:34:19.113 { 00:34:19.113 "name": "BaseBdev2", 00:34:19.113 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:19.113 "is_configured": true, 00:34:19.113 "data_offset": 2048, 00:34:19.113 "data_size": 63488 00:34:19.113 }, 00:34:19.113 { 00:34:19.113 "name": "BaseBdev3", 00:34:19.113 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:19.113 "is_configured": true, 00:34:19.113 "data_offset": 2048, 00:34:19.113 "data_size": 63488 00:34:19.113 }, 00:34:19.113 { 00:34:19.113 "name": "BaseBdev4", 00:34:19.113 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:19.113 "is_configured": true, 00:34:19.113 "data_offset": 2048, 00:34:19.113 "data_size": 63488 00:34:19.113 } 00:34:19.113 ] 00:34:19.113 }' 00:34:19.113 13:54:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:19.113 13:54:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:34:19.113 13:54:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:19.113 13:54:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:34:19.113 13:54:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:34:20.046 13:54:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:34:20.046 13:54:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:20.046 13:54:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:20.046 13:54:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:20.046 13:54:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:20.046 13:54:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:20.046 13:54:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:20.046 13:54:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:20.046 13:54:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:20.046 13:54:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:20.046 13:54:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:20.046 13:54:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:20.046 "name": "raid_bdev1", 00:34:20.046 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:20.046 "strip_size_kb": 64, 00:34:20.046 "state": "online", 00:34:20.046 "raid_level": "raid5f", 00:34:20.046 "superblock": true, 00:34:20.046 "num_base_bdevs": 4, 00:34:20.046 "num_base_bdevs_discovered": 4, 00:34:20.046 "num_base_bdevs_operational": 4, 00:34:20.046 "process": { 00:34:20.046 "type": "rebuild", 00:34:20.046 "target": "spare", 00:34:20.046 "progress": { 00:34:20.046 "blocks": 124800, 00:34:20.046 "percent": 65 00:34:20.046 } 00:34:20.046 }, 00:34:20.046 "base_bdevs_list": [ 00:34:20.046 { 00:34:20.046 "name": "spare", 00:34:20.046 "uuid": "cdf0ebd9-45ac-5c95-8cc7-c198588e1fef", 00:34:20.046 "is_configured": true, 00:34:20.046 "data_offset": 2048, 00:34:20.046 "data_size": 63488 00:34:20.046 }, 00:34:20.046 { 00:34:20.046 "name": "BaseBdev2", 00:34:20.046 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:20.046 "is_configured": true, 00:34:20.046 "data_offset": 2048, 00:34:20.046 "data_size": 63488 00:34:20.046 }, 00:34:20.046 { 00:34:20.046 "name": "BaseBdev3", 00:34:20.046 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:20.046 "is_configured": true, 00:34:20.046 "data_offset": 2048, 00:34:20.046 "data_size": 63488 00:34:20.046 }, 00:34:20.046 { 00:34:20.046 "name": "BaseBdev4", 00:34:20.046 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:20.046 "is_configured": true, 00:34:20.046 "data_offset": 2048, 00:34:20.046 "data_size": 63488 00:34:20.046 } 00:34:20.046 ] 00:34:20.046 }' 00:34:20.046 13:54:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:20.046 13:54:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:34:20.046 13:54:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:20.046 13:54:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:34:20.046 13:54:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:34:20.979 13:54:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:34:20.979 13:54:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:20.979 13:54:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:20.979 13:54:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:20.979 13:54:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:20.979 13:54:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:20.979 13:54:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:20.979 13:54:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:20.979 13:54:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:20.979 13:54:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:20.979 13:54:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:20.979 13:54:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:20.979 "name": "raid_bdev1", 00:34:20.979 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:20.979 "strip_size_kb": 64, 00:34:20.979 "state": "online", 00:34:20.979 "raid_level": "raid5f", 00:34:20.979 "superblock": true, 00:34:20.979 "num_base_bdevs": 4, 00:34:20.979 "num_base_bdevs_discovered": 4, 00:34:20.979 "num_base_bdevs_operational": 4, 00:34:20.979 "process": { 00:34:20.979 "type": "rebuild", 00:34:20.979 "target": "spare", 00:34:20.979 "progress": { 00:34:20.979 "blocks": 145920, 00:34:20.979 "percent": 76 00:34:20.979 } 00:34:20.979 }, 00:34:20.979 "base_bdevs_list": [ 00:34:20.979 { 00:34:20.979 "name": "spare", 00:34:20.979 "uuid": "cdf0ebd9-45ac-5c95-8cc7-c198588e1fef", 00:34:20.979 "is_configured": true, 00:34:20.979 "data_offset": 2048, 00:34:20.979 "data_size": 63488 00:34:20.979 }, 00:34:20.979 { 00:34:20.979 "name": "BaseBdev2", 00:34:20.979 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:20.979 "is_configured": true, 00:34:20.979 "data_offset": 2048, 00:34:20.979 "data_size": 63488 00:34:20.979 }, 00:34:20.979 { 00:34:20.979 "name": "BaseBdev3", 00:34:20.979 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:20.979 "is_configured": true, 00:34:20.979 "data_offset": 2048, 00:34:20.979 "data_size": 63488 00:34:20.979 }, 00:34:20.979 { 00:34:20.979 "name": "BaseBdev4", 00:34:20.979 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:20.979 "is_configured": true, 00:34:20.979 "data_offset": 2048, 00:34:20.979 "data_size": 63488 00:34:20.979 } 00:34:20.979 ] 00:34:20.979 }' 00:34:20.979 13:54:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:20.979 13:54:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:34:20.979 13:54:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:20.979 13:54:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:34:20.979 13:54:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:34:22.348 13:54:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:34:22.348 13:54:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:22.348 13:54:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:22.348 13:54:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:22.348 13:54:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:22.348 13:54:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:22.348 13:54:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:22.348 13:54:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:22.348 13:54:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:22.348 13:54:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:22.348 13:54:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:22.348 13:54:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:22.348 "name": "raid_bdev1", 00:34:22.348 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:22.348 "strip_size_kb": 64, 00:34:22.348 "state": "online", 00:34:22.348 "raid_level": "raid5f", 00:34:22.348 "superblock": true, 00:34:22.348 "num_base_bdevs": 4, 00:34:22.348 "num_base_bdevs_discovered": 4, 00:34:22.348 "num_base_bdevs_operational": 4, 00:34:22.348 "process": { 00:34:22.348 "type": "rebuild", 00:34:22.348 "target": "spare", 00:34:22.348 "progress": { 00:34:22.348 "blocks": 167040, 00:34:22.348 "percent": 87 00:34:22.348 } 00:34:22.348 }, 00:34:22.348 "base_bdevs_list": [ 00:34:22.348 { 00:34:22.348 "name": "spare", 00:34:22.348 "uuid": "cdf0ebd9-45ac-5c95-8cc7-c198588e1fef", 00:34:22.348 "is_configured": true, 00:34:22.348 "data_offset": 2048, 00:34:22.348 "data_size": 63488 00:34:22.348 }, 00:34:22.348 { 00:34:22.348 "name": "BaseBdev2", 00:34:22.348 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:22.348 "is_configured": true, 00:34:22.348 "data_offset": 2048, 00:34:22.348 "data_size": 63488 00:34:22.348 }, 00:34:22.348 { 00:34:22.348 "name": "BaseBdev3", 00:34:22.348 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:22.348 "is_configured": true, 00:34:22.348 "data_offset": 2048, 00:34:22.348 "data_size": 63488 00:34:22.348 }, 00:34:22.348 { 00:34:22.348 "name": "BaseBdev4", 00:34:22.348 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:22.348 "is_configured": true, 00:34:22.348 "data_offset": 2048, 00:34:22.348 "data_size": 63488 00:34:22.348 } 00:34:22.348 ] 00:34:22.348 }' 00:34:22.348 13:54:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:22.348 13:54:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:34:22.348 13:54:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:22.348 13:54:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:34:22.348 13:54:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:34:23.282 13:54:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:34:23.282 13:54:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:23.282 13:54:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:23.282 13:54:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:23.282 13:54:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:23.282 13:54:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:23.282 13:54:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:23.282 13:54:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:23.282 13:54:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:23.282 13:54:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:23.282 13:54:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:23.282 13:54:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:23.282 "name": "raid_bdev1", 00:34:23.282 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:23.282 "strip_size_kb": 64, 00:34:23.282 "state": "online", 00:34:23.282 "raid_level": "raid5f", 00:34:23.282 "superblock": true, 00:34:23.282 "num_base_bdevs": 4, 00:34:23.282 "num_base_bdevs_discovered": 4, 00:34:23.282 "num_base_bdevs_operational": 4, 00:34:23.282 "process": { 00:34:23.282 "type": "rebuild", 00:34:23.282 "target": "spare", 00:34:23.282 "progress": { 00:34:23.282 "blocks": 188160, 00:34:23.282 "percent": 98 00:34:23.282 } 00:34:23.282 }, 00:34:23.282 "base_bdevs_list": [ 00:34:23.282 { 00:34:23.282 "name": "spare", 00:34:23.282 "uuid": "cdf0ebd9-45ac-5c95-8cc7-c198588e1fef", 00:34:23.282 "is_configured": true, 00:34:23.282 "data_offset": 2048, 00:34:23.282 "data_size": 63488 00:34:23.282 }, 00:34:23.282 { 00:34:23.282 "name": "BaseBdev2", 00:34:23.282 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:23.282 "is_configured": true, 00:34:23.282 "data_offset": 2048, 00:34:23.282 "data_size": 63488 00:34:23.282 }, 00:34:23.282 { 00:34:23.282 "name": "BaseBdev3", 00:34:23.282 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:23.282 "is_configured": true, 00:34:23.282 "data_offset": 2048, 00:34:23.282 "data_size": 63488 00:34:23.282 }, 00:34:23.282 { 00:34:23.282 "name": "BaseBdev4", 00:34:23.282 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:23.282 "is_configured": true, 00:34:23.282 "data_offset": 2048, 00:34:23.282 "data_size": 63488 00:34:23.282 } 00:34:23.282 ] 00:34:23.282 }' 00:34:23.282 13:54:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:23.282 13:54:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:34:23.282 13:54:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:23.282 13:54:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:34:23.282 13:54:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:34:23.282 [2024-11-20 13:54:30.305278] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:34:23.282 [2024-11-20 13:54:30.305354] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:34:23.282 [2024-11-20 13:54:30.305497] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:34:24.265 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:34:24.265 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:24.265 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:24.265 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:24.265 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:24.265 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:24.265 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:24.265 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:24.265 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:24.265 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:24.265 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:24.265 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:24.265 "name": "raid_bdev1", 00:34:24.265 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:24.265 "strip_size_kb": 64, 00:34:24.265 "state": "online", 00:34:24.265 "raid_level": "raid5f", 00:34:24.265 "superblock": true, 00:34:24.265 "num_base_bdevs": 4, 00:34:24.265 "num_base_bdevs_discovered": 4, 00:34:24.265 "num_base_bdevs_operational": 4, 00:34:24.265 "base_bdevs_list": [ 00:34:24.265 { 00:34:24.265 "name": "spare", 00:34:24.265 "uuid": "cdf0ebd9-45ac-5c95-8cc7-c198588e1fef", 00:34:24.265 "is_configured": true, 00:34:24.265 "data_offset": 2048, 00:34:24.265 "data_size": 63488 00:34:24.265 }, 00:34:24.265 { 00:34:24.265 "name": "BaseBdev2", 00:34:24.265 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:24.265 "is_configured": true, 00:34:24.265 "data_offset": 2048, 00:34:24.265 "data_size": 63488 00:34:24.265 }, 00:34:24.265 { 00:34:24.265 "name": "BaseBdev3", 00:34:24.265 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:24.265 "is_configured": true, 00:34:24.265 "data_offset": 2048, 00:34:24.265 "data_size": 63488 00:34:24.265 }, 00:34:24.265 { 00:34:24.265 "name": "BaseBdev4", 00:34:24.265 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:24.265 "is_configured": true, 00:34:24.265 "data_offset": 2048, 00:34:24.265 "data_size": 63488 00:34:24.265 } 00:34:24.265 ] 00:34:24.265 }' 00:34:24.265 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:24.265 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:34:24.265 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:24.523 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:34:24.523 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:34:24.523 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:34:24.523 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:24.523 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:34:24.523 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:34:24.523 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:24.523 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:24.523 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:24.523 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:24.523 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:24.523 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:24.523 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:24.523 "name": "raid_bdev1", 00:34:24.523 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:24.523 "strip_size_kb": 64, 00:34:24.523 "state": "online", 00:34:24.523 "raid_level": "raid5f", 00:34:24.523 "superblock": true, 00:34:24.523 "num_base_bdevs": 4, 00:34:24.523 "num_base_bdevs_discovered": 4, 00:34:24.523 "num_base_bdevs_operational": 4, 00:34:24.523 "base_bdevs_list": [ 00:34:24.523 { 00:34:24.523 "name": "spare", 00:34:24.523 "uuid": "cdf0ebd9-45ac-5c95-8cc7-c198588e1fef", 00:34:24.523 "is_configured": true, 00:34:24.523 "data_offset": 2048, 00:34:24.523 "data_size": 63488 00:34:24.523 }, 00:34:24.523 { 00:34:24.523 "name": "BaseBdev2", 00:34:24.523 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:24.523 "is_configured": true, 00:34:24.523 "data_offset": 2048, 00:34:24.523 "data_size": 63488 00:34:24.523 }, 00:34:24.523 { 00:34:24.523 "name": "BaseBdev3", 00:34:24.523 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:24.523 "is_configured": true, 00:34:24.523 "data_offset": 2048, 00:34:24.523 "data_size": 63488 00:34:24.523 }, 00:34:24.523 { 00:34:24.523 "name": "BaseBdev4", 00:34:24.523 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:24.523 "is_configured": true, 00:34:24.523 "data_offset": 2048, 00:34:24.523 "data_size": 63488 00:34:24.523 } 00:34:24.523 ] 00:34:24.523 }' 00:34:24.523 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:24.523 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:34:24.523 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:24.523 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:34:24.523 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:34:24.523 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:24.523 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:24.523 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:34:24.523 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:34:24.523 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:34:24.523 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:24.523 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:24.523 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:24.524 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:24.524 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:24.524 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:24.524 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:24.524 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:24.524 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:24.524 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:24.524 "name": "raid_bdev1", 00:34:24.524 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:24.524 "strip_size_kb": 64, 00:34:24.524 "state": "online", 00:34:24.524 "raid_level": "raid5f", 00:34:24.524 "superblock": true, 00:34:24.524 "num_base_bdevs": 4, 00:34:24.524 "num_base_bdevs_discovered": 4, 00:34:24.524 "num_base_bdevs_operational": 4, 00:34:24.524 "base_bdevs_list": [ 00:34:24.524 { 00:34:24.524 "name": "spare", 00:34:24.524 "uuid": "cdf0ebd9-45ac-5c95-8cc7-c198588e1fef", 00:34:24.524 "is_configured": true, 00:34:24.524 "data_offset": 2048, 00:34:24.524 "data_size": 63488 00:34:24.524 }, 00:34:24.524 { 00:34:24.524 "name": "BaseBdev2", 00:34:24.524 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:24.524 "is_configured": true, 00:34:24.524 "data_offset": 2048, 00:34:24.524 "data_size": 63488 00:34:24.524 }, 00:34:24.524 { 00:34:24.524 "name": "BaseBdev3", 00:34:24.524 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:24.524 "is_configured": true, 00:34:24.524 "data_offset": 2048, 00:34:24.524 "data_size": 63488 00:34:24.524 }, 00:34:24.524 { 00:34:24.524 "name": "BaseBdev4", 00:34:24.524 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:24.524 "is_configured": true, 00:34:24.524 "data_offset": 2048, 00:34:24.524 "data_size": 63488 00:34:24.524 } 00:34:24.524 ] 00:34:24.524 }' 00:34:24.524 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:24.524 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:24.782 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:34:24.782 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:24.782 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:24.782 [2024-11-20 13:54:31.763229] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:34:24.782 [2024-11-20 13:54:31.763264] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:34:24.782 [2024-11-20 13:54:31.763347] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:34:24.782 [2024-11-20 13:54:31.763442] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:34:24.782 [2024-11-20 13:54:31.763453] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:34:24.782 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:24.782 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:24.782 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:24.782 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:34:24.782 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:24.782 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:24.782 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:34:24.782 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:34:24.782 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:34:24.782 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:34:24.783 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:34:24.783 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:34:24.783 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:34:24.783 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:34:24.783 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:34:24.783 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:34:24.783 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:34:24.783 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:34:24.783 13:54:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:34:25.040 /dev/nbd0 00:34:25.040 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:34:25.040 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:34:25.040 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:34:25.040 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:34:25.040 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:34:25.040 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:34:25.040 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:34:25.040 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:34:25.040 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:34:25.040 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:34:25.040 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:34:25.040 1+0 records in 00:34:25.040 1+0 records out 00:34:25.040 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000215407 s, 19.0 MB/s 00:34:25.040 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:34:25.040 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:34:25.040 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:34:25.040 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:34:25.040 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:34:25.040 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:34:25.040 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:34:25.040 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:34:25.299 /dev/nbd1 00:34:25.299 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:34:25.299 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:34:25.299 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:34:25.299 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:34:25.299 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:34:25.299 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:34:25.299 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:34:25.299 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:34:25.299 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:34:25.299 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:34:25.299 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:34:25.299 1+0 records in 00:34:25.299 1+0 records out 00:34:25.299 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000279044 s, 14.7 MB/s 00:34:25.299 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:34:25.299 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:34:25.299 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:34:25.299 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:34:25.299 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:34:25.299 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:34:25.299 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:34:25.299 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:34:25.556 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:34:25.556 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:34:25.556 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:34:25.556 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:34:25.556 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:34:25.556 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:34:25.556 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:34:25.556 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:34:25.556 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:34:25.556 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:34:25.556 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:34:25.556 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:34:25.556 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:34:25.556 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:34:25.556 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:34:25.556 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:34:25.556 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:34:25.813 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:34:25.813 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:34:25.813 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:34:25.813 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:34:25.813 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:34:25.813 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:34:25.813 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:34:25.813 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:34:25.813 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:34:25.813 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:34:25.813 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:25.813 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:25.814 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:25.814 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:34:25.814 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:25.814 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:25.814 [2024-11-20 13:54:32.848098] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:34:25.814 [2024-11-20 13:54:32.848152] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:34:25.814 [2024-11-20 13:54:32.848175] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:34:25.814 [2024-11-20 13:54:32.848184] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:34:25.814 [2024-11-20 13:54:32.850304] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:34:25.814 [2024-11-20 13:54:32.850462] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:34:25.814 [2024-11-20 13:54:32.850567] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:34:25.814 [2024-11-20 13:54:32.850617] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:34:25.814 [2024-11-20 13:54:32.850745] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:34:25.814 [2024-11-20 13:54:32.850825] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:34:25.814 [2024-11-20 13:54:32.850892] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:34:25.814 spare 00:34:25.814 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:25.814 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:34:25.814 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:25.814 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:26.071 [2024-11-20 13:54:32.950974] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:34:26.071 [2024-11-20 13:54:32.951013] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:34:26.071 [2024-11-20 13:54:32.951289] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000491d0 00:34:26.071 [2024-11-20 13:54:32.955140] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:34:26.071 [2024-11-20 13:54:32.955157] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:34:26.071 [2024-11-20 13:54:32.955333] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:34:26.071 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:26.071 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:34:26.071 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:26.071 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:26.071 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:34:26.071 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:34:26.071 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:34:26.071 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:26.071 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:26.071 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:26.071 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:26.071 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:26.071 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:26.071 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:26.071 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:26.071 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:26.071 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:26.071 "name": "raid_bdev1", 00:34:26.071 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:26.071 "strip_size_kb": 64, 00:34:26.071 "state": "online", 00:34:26.071 "raid_level": "raid5f", 00:34:26.071 "superblock": true, 00:34:26.071 "num_base_bdevs": 4, 00:34:26.071 "num_base_bdevs_discovered": 4, 00:34:26.071 "num_base_bdevs_operational": 4, 00:34:26.071 "base_bdevs_list": [ 00:34:26.071 { 00:34:26.071 "name": "spare", 00:34:26.071 "uuid": "cdf0ebd9-45ac-5c95-8cc7-c198588e1fef", 00:34:26.071 "is_configured": true, 00:34:26.071 "data_offset": 2048, 00:34:26.071 "data_size": 63488 00:34:26.071 }, 00:34:26.071 { 00:34:26.071 "name": "BaseBdev2", 00:34:26.071 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:26.071 "is_configured": true, 00:34:26.071 "data_offset": 2048, 00:34:26.071 "data_size": 63488 00:34:26.071 }, 00:34:26.071 { 00:34:26.071 "name": "BaseBdev3", 00:34:26.071 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:26.071 "is_configured": true, 00:34:26.071 "data_offset": 2048, 00:34:26.071 "data_size": 63488 00:34:26.071 }, 00:34:26.071 { 00:34:26.071 "name": "BaseBdev4", 00:34:26.071 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:26.071 "is_configured": true, 00:34:26.071 "data_offset": 2048, 00:34:26.071 "data_size": 63488 00:34:26.071 } 00:34:26.071 ] 00:34:26.071 }' 00:34:26.071 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:26.071 13:54:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:26.329 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:34:26.329 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:26.329 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:34:26.329 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:34:26.329 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:26.329 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:26.329 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:26.329 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:26.329 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:26.329 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:26.329 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:26.329 "name": "raid_bdev1", 00:34:26.329 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:26.329 "strip_size_kb": 64, 00:34:26.329 "state": "online", 00:34:26.329 "raid_level": "raid5f", 00:34:26.329 "superblock": true, 00:34:26.329 "num_base_bdevs": 4, 00:34:26.329 "num_base_bdevs_discovered": 4, 00:34:26.329 "num_base_bdevs_operational": 4, 00:34:26.329 "base_bdevs_list": [ 00:34:26.329 { 00:34:26.329 "name": "spare", 00:34:26.329 "uuid": "cdf0ebd9-45ac-5c95-8cc7-c198588e1fef", 00:34:26.329 "is_configured": true, 00:34:26.329 "data_offset": 2048, 00:34:26.330 "data_size": 63488 00:34:26.330 }, 00:34:26.330 { 00:34:26.330 "name": "BaseBdev2", 00:34:26.330 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:26.330 "is_configured": true, 00:34:26.330 "data_offset": 2048, 00:34:26.330 "data_size": 63488 00:34:26.330 }, 00:34:26.330 { 00:34:26.330 "name": "BaseBdev3", 00:34:26.330 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:26.330 "is_configured": true, 00:34:26.330 "data_offset": 2048, 00:34:26.330 "data_size": 63488 00:34:26.330 }, 00:34:26.330 { 00:34:26.330 "name": "BaseBdev4", 00:34:26.330 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:26.330 "is_configured": true, 00:34:26.330 "data_offset": 2048, 00:34:26.330 "data_size": 63488 00:34:26.330 } 00:34:26.330 ] 00:34:26.330 }' 00:34:26.330 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:26.330 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:34:26.330 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:26.330 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:34:26.330 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:26.330 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:34:26.330 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:26.330 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:26.330 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:26.588 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:34:26.588 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:34:26.588 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:26.588 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:26.588 [2024-11-20 13:54:33.404337] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:34:26.588 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:26.588 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:34:26.588 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:26.588 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:26.588 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:34:26.588 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:34:26.588 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:34:26.588 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:26.588 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:26.588 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:26.588 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:26.588 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:26.588 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:26.588 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:26.588 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:26.588 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:26.588 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:26.588 "name": "raid_bdev1", 00:34:26.588 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:26.588 "strip_size_kb": 64, 00:34:26.588 "state": "online", 00:34:26.588 "raid_level": "raid5f", 00:34:26.588 "superblock": true, 00:34:26.588 "num_base_bdevs": 4, 00:34:26.588 "num_base_bdevs_discovered": 3, 00:34:26.588 "num_base_bdevs_operational": 3, 00:34:26.588 "base_bdevs_list": [ 00:34:26.588 { 00:34:26.588 "name": null, 00:34:26.588 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:26.588 "is_configured": false, 00:34:26.588 "data_offset": 0, 00:34:26.588 "data_size": 63488 00:34:26.588 }, 00:34:26.588 { 00:34:26.588 "name": "BaseBdev2", 00:34:26.588 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:26.588 "is_configured": true, 00:34:26.588 "data_offset": 2048, 00:34:26.588 "data_size": 63488 00:34:26.588 }, 00:34:26.588 { 00:34:26.588 "name": "BaseBdev3", 00:34:26.588 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:26.588 "is_configured": true, 00:34:26.588 "data_offset": 2048, 00:34:26.588 "data_size": 63488 00:34:26.588 }, 00:34:26.588 { 00:34:26.588 "name": "BaseBdev4", 00:34:26.588 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:26.588 "is_configured": true, 00:34:26.588 "data_offset": 2048, 00:34:26.588 "data_size": 63488 00:34:26.588 } 00:34:26.588 ] 00:34:26.588 }' 00:34:26.588 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:26.588 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:26.846 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:34:26.846 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:26.846 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:26.846 [2024-11-20 13:54:33.724431] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:34:26.846 [2024-11-20 13:54:33.724746] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:34:26.846 [2024-11-20 13:54:33.724773] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:34:26.846 [2024-11-20 13:54:33.724809] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:34:26.846 [2024-11-20 13:54:33.732902] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000492a0 00:34:26.846 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:26.846 13:54:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:34:26.846 [2024-11-20 13:54:33.738606] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:34:27.779 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:27.779 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:27.779 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:27.779 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:27.779 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:27.779 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:27.779 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:27.779 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:27.779 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:27.779 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:27.779 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:27.779 "name": "raid_bdev1", 00:34:27.779 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:27.779 "strip_size_kb": 64, 00:34:27.779 "state": "online", 00:34:27.779 "raid_level": "raid5f", 00:34:27.779 "superblock": true, 00:34:27.779 "num_base_bdevs": 4, 00:34:27.779 "num_base_bdevs_discovered": 4, 00:34:27.779 "num_base_bdevs_operational": 4, 00:34:27.779 "process": { 00:34:27.779 "type": "rebuild", 00:34:27.779 "target": "spare", 00:34:27.779 "progress": { 00:34:27.779 "blocks": 19200, 00:34:27.779 "percent": 10 00:34:27.779 } 00:34:27.779 }, 00:34:27.779 "base_bdevs_list": [ 00:34:27.779 { 00:34:27.779 "name": "spare", 00:34:27.779 "uuid": "cdf0ebd9-45ac-5c95-8cc7-c198588e1fef", 00:34:27.779 "is_configured": true, 00:34:27.779 "data_offset": 2048, 00:34:27.779 "data_size": 63488 00:34:27.779 }, 00:34:27.779 { 00:34:27.779 "name": "BaseBdev2", 00:34:27.779 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:27.779 "is_configured": true, 00:34:27.779 "data_offset": 2048, 00:34:27.779 "data_size": 63488 00:34:27.779 }, 00:34:27.779 { 00:34:27.779 "name": "BaseBdev3", 00:34:27.779 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:27.779 "is_configured": true, 00:34:27.779 "data_offset": 2048, 00:34:27.779 "data_size": 63488 00:34:27.779 }, 00:34:27.779 { 00:34:27.779 "name": "BaseBdev4", 00:34:27.779 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:27.779 "is_configured": true, 00:34:27.779 "data_offset": 2048, 00:34:27.779 "data_size": 63488 00:34:27.779 } 00:34:27.779 ] 00:34:27.779 }' 00:34:27.779 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:27.779 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:34:27.779 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:28.037 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:34:28.037 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:34:28.037 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:28.037 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:28.037 [2024-11-20 13:54:34.839443] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:34:28.037 [2024-11-20 13:54:34.846942] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:34:28.037 [2024-11-20 13:54:34.847015] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:34:28.037 [2024-11-20 13:54:34.847033] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:34:28.037 [2024-11-20 13:54:34.847041] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:34:28.037 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:28.037 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:34:28.037 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:28.037 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:28.037 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:34:28.037 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:34:28.037 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:34:28.037 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:28.037 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:28.037 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:28.037 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:28.037 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:28.037 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:28.037 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:28.037 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:28.037 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:28.037 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:28.037 "name": "raid_bdev1", 00:34:28.037 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:28.037 "strip_size_kb": 64, 00:34:28.037 "state": "online", 00:34:28.037 "raid_level": "raid5f", 00:34:28.037 "superblock": true, 00:34:28.037 "num_base_bdevs": 4, 00:34:28.037 "num_base_bdevs_discovered": 3, 00:34:28.037 "num_base_bdevs_operational": 3, 00:34:28.037 "base_bdevs_list": [ 00:34:28.037 { 00:34:28.037 "name": null, 00:34:28.037 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:28.037 "is_configured": false, 00:34:28.037 "data_offset": 0, 00:34:28.037 "data_size": 63488 00:34:28.037 }, 00:34:28.037 { 00:34:28.037 "name": "BaseBdev2", 00:34:28.037 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:28.037 "is_configured": true, 00:34:28.037 "data_offset": 2048, 00:34:28.037 "data_size": 63488 00:34:28.037 }, 00:34:28.037 { 00:34:28.037 "name": "BaseBdev3", 00:34:28.037 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:28.037 "is_configured": true, 00:34:28.037 "data_offset": 2048, 00:34:28.037 "data_size": 63488 00:34:28.037 }, 00:34:28.037 { 00:34:28.037 "name": "BaseBdev4", 00:34:28.037 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:28.037 "is_configured": true, 00:34:28.037 "data_offset": 2048, 00:34:28.037 "data_size": 63488 00:34:28.037 } 00:34:28.037 ] 00:34:28.037 }' 00:34:28.037 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:28.037 13:54:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:28.296 13:54:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:34:28.296 13:54:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:28.296 13:54:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:28.296 [2024-11-20 13:54:35.180358] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:34:28.296 [2024-11-20 13:54:35.180434] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:34:28.296 [2024-11-20 13:54:35.180459] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c380 00:34:28.296 [2024-11-20 13:54:35.180470] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:34:28.296 [2024-11-20 13:54:35.180960] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:34:28.296 [2024-11-20 13:54:35.180975] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:34:28.296 [2024-11-20 13:54:35.181088] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:34:28.296 [2024-11-20 13:54:35.181102] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:34:28.296 [2024-11-20 13:54:35.181111] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:34:28.296 [2024-11-20 13:54:35.181131] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:34:28.296 [2024-11-20 13:54:35.189293] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000049370 00:34:28.296 spare 00:34:28.296 13:54:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:28.296 13:54:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:34:28.296 [2024-11-20 13:54:35.194695] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:34:29.230 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:29.230 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:29.230 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:29.230 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:29.230 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:29.230 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:29.230 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:29.230 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:29.230 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:29.230 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:29.230 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:29.230 "name": "raid_bdev1", 00:34:29.230 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:29.230 "strip_size_kb": 64, 00:34:29.230 "state": "online", 00:34:29.230 "raid_level": "raid5f", 00:34:29.230 "superblock": true, 00:34:29.230 "num_base_bdevs": 4, 00:34:29.230 "num_base_bdevs_discovered": 4, 00:34:29.230 "num_base_bdevs_operational": 4, 00:34:29.230 "process": { 00:34:29.230 "type": "rebuild", 00:34:29.230 "target": "spare", 00:34:29.230 "progress": { 00:34:29.230 "blocks": 19200, 00:34:29.230 "percent": 10 00:34:29.230 } 00:34:29.230 }, 00:34:29.230 "base_bdevs_list": [ 00:34:29.230 { 00:34:29.230 "name": "spare", 00:34:29.230 "uuid": "cdf0ebd9-45ac-5c95-8cc7-c198588e1fef", 00:34:29.230 "is_configured": true, 00:34:29.230 "data_offset": 2048, 00:34:29.230 "data_size": 63488 00:34:29.230 }, 00:34:29.230 { 00:34:29.231 "name": "BaseBdev2", 00:34:29.231 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:29.231 "is_configured": true, 00:34:29.231 "data_offset": 2048, 00:34:29.231 "data_size": 63488 00:34:29.231 }, 00:34:29.231 { 00:34:29.231 "name": "BaseBdev3", 00:34:29.231 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:29.231 "is_configured": true, 00:34:29.231 "data_offset": 2048, 00:34:29.231 "data_size": 63488 00:34:29.231 }, 00:34:29.231 { 00:34:29.231 "name": "BaseBdev4", 00:34:29.231 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:29.231 "is_configured": true, 00:34:29.231 "data_offset": 2048, 00:34:29.231 "data_size": 63488 00:34:29.231 } 00:34:29.231 ] 00:34:29.231 }' 00:34:29.231 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:29.231 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:34:29.231 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:29.231 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:34:29.231 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:34:29.231 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:29.231 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:29.231 [2024-11-20 13:54:36.283624] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:34:29.488 [2024-11-20 13:54:36.303330] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:34:29.488 [2024-11-20 13:54:36.303381] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:34:29.488 [2024-11-20 13:54:36.303398] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:34:29.488 [2024-11-20 13:54:36.303404] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:34:29.488 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:29.488 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:34:29.488 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:29.488 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:29.488 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:34:29.488 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:34:29.488 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:34:29.488 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:29.488 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:29.488 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:29.488 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:29.488 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:29.488 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:29.488 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:29.488 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:29.488 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:29.488 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:29.488 "name": "raid_bdev1", 00:34:29.488 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:29.488 "strip_size_kb": 64, 00:34:29.488 "state": "online", 00:34:29.488 "raid_level": "raid5f", 00:34:29.488 "superblock": true, 00:34:29.488 "num_base_bdevs": 4, 00:34:29.488 "num_base_bdevs_discovered": 3, 00:34:29.488 "num_base_bdevs_operational": 3, 00:34:29.488 "base_bdevs_list": [ 00:34:29.488 { 00:34:29.488 "name": null, 00:34:29.488 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:29.488 "is_configured": false, 00:34:29.488 "data_offset": 0, 00:34:29.488 "data_size": 63488 00:34:29.488 }, 00:34:29.488 { 00:34:29.488 "name": "BaseBdev2", 00:34:29.488 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:29.488 "is_configured": true, 00:34:29.488 "data_offset": 2048, 00:34:29.488 "data_size": 63488 00:34:29.488 }, 00:34:29.488 { 00:34:29.488 "name": "BaseBdev3", 00:34:29.488 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:29.488 "is_configured": true, 00:34:29.488 "data_offset": 2048, 00:34:29.488 "data_size": 63488 00:34:29.488 }, 00:34:29.488 { 00:34:29.488 "name": "BaseBdev4", 00:34:29.488 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:29.488 "is_configured": true, 00:34:29.489 "data_offset": 2048, 00:34:29.489 "data_size": 63488 00:34:29.489 } 00:34:29.489 ] 00:34:29.489 }' 00:34:29.489 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:29.489 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:29.746 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:34:29.746 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:29.746 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:34:29.746 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:34:29.746 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:29.746 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:29.746 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:29.746 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:29.746 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:29.746 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:29.746 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:29.746 "name": "raid_bdev1", 00:34:29.746 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:29.746 "strip_size_kb": 64, 00:34:29.746 "state": "online", 00:34:29.746 "raid_level": "raid5f", 00:34:29.746 "superblock": true, 00:34:29.746 "num_base_bdevs": 4, 00:34:29.746 "num_base_bdevs_discovered": 3, 00:34:29.746 "num_base_bdevs_operational": 3, 00:34:29.746 "base_bdevs_list": [ 00:34:29.746 { 00:34:29.746 "name": null, 00:34:29.746 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:29.746 "is_configured": false, 00:34:29.746 "data_offset": 0, 00:34:29.746 "data_size": 63488 00:34:29.746 }, 00:34:29.746 { 00:34:29.746 "name": "BaseBdev2", 00:34:29.746 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:29.746 "is_configured": true, 00:34:29.746 "data_offset": 2048, 00:34:29.746 "data_size": 63488 00:34:29.746 }, 00:34:29.746 { 00:34:29.746 "name": "BaseBdev3", 00:34:29.746 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:29.746 "is_configured": true, 00:34:29.746 "data_offset": 2048, 00:34:29.746 "data_size": 63488 00:34:29.746 }, 00:34:29.746 { 00:34:29.746 "name": "BaseBdev4", 00:34:29.746 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:29.746 "is_configured": true, 00:34:29.746 "data_offset": 2048, 00:34:29.746 "data_size": 63488 00:34:29.746 } 00:34:29.746 ] 00:34:29.746 }' 00:34:29.746 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:29.746 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:34:29.746 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:29.746 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:34:29.746 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:34:29.746 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:29.746 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:29.746 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:29.747 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:34:29.747 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:29.747 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:29.747 [2024-11-20 13:54:36.756512] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:34:29.747 [2024-11-20 13:54:36.756571] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:34:29.747 [2024-11-20 13:54:36.756593] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c980 00:34:29.747 [2024-11-20 13:54:36.756601] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:34:29.747 [2024-11-20 13:54:36.757071] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:34:29.747 [2024-11-20 13:54:36.757085] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:34:29.747 [2024-11-20 13:54:36.757161] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:34:29.747 [2024-11-20 13:54:36.757173] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:34:29.747 [2024-11-20 13:54:36.757181] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:34:29.747 [2024-11-20 13:54:36.757191] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:34:29.747 BaseBdev1 00:34:29.747 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:29.747 13:54:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:34:31.121 13:54:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:34:31.121 13:54:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:31.121 13:54:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:31.121 13:54:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:34:31.121 13:54:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:34:31.121 13:54:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:34:31.121 13:54:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:31.121 13:54:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:31.121 13:54:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:31.121 13:54:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:31.121 13:54:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:31.121 13:54:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:31.121 13:54:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:31.121 13:54:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:31.121 13:54:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:31.121 13:54:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:31.121 "name": "raid_bdev1", 00:34:31.121 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:31.121 "strip_size_kb": 64, 00:34:31.121 "state": "online", 00:34:31.121 "raid_level": "raid5f", 00:34:31.121 "superblock": true, 00:34:31.121 "num_base_bdevs": 4, 00:34:31.121 "num_base_bdevs_discovered": 3, 00:34:31.121 "num_base_bdevs_operational": 3, 00:34:31.121 "base_bdevs_list": [ 00:34:31.121 { 00:34:31.121 "name": null, 00:34:31.121 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:31.121 "is_configured": false, 00:34:31.121 "data_offset": 0, 00:34:31.121 "data_size": 63488 00:34:31.121 }, 00:34:31.121 { 00:34:31.121 "name": "BaseBdev2", 00:34:31.121 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:31.121 "is_configured": true, 00:34:31.121 "data_offset": 2048, 00:34:31.121 "data_size": 63488 00:34:31.121 }, 00:34:31.121 { 00:34:31.121 "name": "BaseBdev3", 00:34:31.121 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:31.121 "is_configured": true, 00:34:31.121 "data_offset": 2048, 00:34:31.121 "data_size": 63488 00:34:31.121 }, 00:34:31.121 { 00:34:31.121 "name": "BaseBdev4", 00:34:31.121 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:31.121 "is_configured": true, 00:34:31.121 "data_offset": 2048, 00:34:31.121 "data_size": 63488 00:34:31.121 } 00:34:31.121 ] 00:34:31.121 }' 00:34:31.121 13:54:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:31.121 13:54:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:31.121 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:34:31.122 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:31.122 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:34:31.122 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:34:31.122 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:31.122 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:31.122 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:31.122 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:31.122 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:31.122 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:31.122 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:31.122 "name": "raid_bdev1", 00:34:31.122 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:31.122 "strip_size_kb": 64, 00:34:31.122 "state": "online", 00:34:31.122 "raid_level": "raid5f", 00:34:31.122 "superblock": true, 00:34:31.122 "num_base_bdevs": 4, 00:34:31.122 "num_base_bdevs_discovered": 3, 00:34:31.122 "num_base_bdevs_operational": 3, 00:34:31.122 "base_bdevs_list": [ 00:34:31.122 { 00:34:31.122 "name": null, 00:34:31.122 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:31.122 "is_configured": false, 00:34:31.122 "data_offset": 0, 00:34:31.122 "data_size": 63488 00:34:31.122 }, 00:34:31.122 { 00:34:31.122 "name": "BaseBdev2", 00:34:31.122 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:31.122 "is_configured": true, 00:34:31.122 "data_offset": 2048, 00:34:31.122 "data_size": 63488 00:34:31.122 }, 00:34:31.122 { 00:34:31.122 "name": "BaseBdev3", 00:34:31.122 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:31.122 "is_configured": true, 00:34:31.122 "data_offset": 2048, 00:34:31.122 "data_size": 63488 00:34:31.122 }, 00:34:31.122 { 00:34:31.122 "name": "BaseBdev4", 00:34:31.122 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:31.122 "is_configured": true, 00:34:31.122 "data_offset": 2048, 00:34:31.122 "data_size": 63488 00:34:31.122 } 00:34:31.122 ] 00:34:31.122 }' 00:34:31.122 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:31.122 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:34:31.122 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:31.380 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:34:31.380 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:34:31.380 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:34:31.380 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:34:31.380 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:34:31.380 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:34:31.380 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:34:31.380 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:34:31.380 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:34:31.380 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:31.380 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:31.380 [2024-11-20 13:54:38.192818] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:34:31.380 [2024-11-20 13:54:38.192991] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:34:31.380 [2024-11-20 13:54:38.193005] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:34:31.380 request: 00:34:31.380 { 00:34:31.380 "base_bdev": "BaseBdev1", 00:34:31.380 "raid_bdev": "raid_bdev1", 00:34:31.380 "method": "bdev_raid_add_base_bdev", 00:34:31.380 "req_id": 1 00:34:31.380 } 00:34:31.380 Got JSON-RPC error response 00:34:31.380 response: 00:34:31.380 { 00:34:31.380 "code": -22, 00:34:31.380 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:34:31.380 } 00:34:31.380 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:34:31.380 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:34:31.380 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:34:31.380 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:34:31.380 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:34:31.380 13:54:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:34:32.316 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:34:32.316 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:32.316 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:32.316 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:34:32.316 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:34:32.316 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:34:32.316 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:32.316 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:32.316 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:32.316 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:32.316 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:32.316 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:32.316 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:32.316 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:32.316 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:32.316 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:32.316 "name": "raid_bdev1", 00:34:32.316 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:32.316 "strip_size_kb": 64, 00:34:32.316 "state": "online", 00:34:32.316 "raid_level": "raid5f", 00:34:32.316 "superblock": true, 00:34:32.316 "num_base_bdevs": 4, 00:34:32.316 "num_base_bdevs_discovered": 3, 00:34:32.316 "num_base_bdevs_operational": 3, 00:34:32.316 "base_bdevs_list": [ 00:34:32.316 { 00:34:32.316 "name": null, 00:34:32.316 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:32.316 "is_configured": false, 00:34:32.316 "data_offset": 0, 00:34:32.316 "data_size": 63488 00:34:32.316 }, 00:34:32.316 { 00:34:32.316 "name": "BaseBdev2", 00:34:32.316 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:32.316 "is_configured": true, 00:34:32.316 "data_offset": 2048, 00:34:32.316 "data_size": 63488 00:34:32.316 }, 00:34:32.316 { 00:34:32.316 "name": "BaseBdev3", 00:34:32.316 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:32.316 "is_configured": true, 00:34:32.316 "data_offset": 2048, 00:34:32.316 "data_size": 63488 00:34:32.316 }, 00:34:32.316 { 00:34:32.316 "name": "BaseBdev4", 00:34:32.316 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:32.316 "is_configured": true, 00:34:32.316 "data_offset": 2048, 00:34:32.316 "data_size": 63488 00:34:32.316 } 00:34:32.316 ] 00:34:32.316 }' 00:34:32.316 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:32.316 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:32.575 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:34:32.575 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:32.575 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:34:32.575 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:34:32.575 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:32.575 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:32.575 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:32.575 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:32.575 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:32.575 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:32.575 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:32.575 "name": "raid_bdev1", 00:34:32.575 "uuid": "a6468f48-51be-4405-bea7-d9f5c10eea55", 00:34:32.575 "strip_size_kb": 64, 00:34:32.575 "state": "online", 00:34:32.575 "raid_level": "raid5f", 00:34:32.575 "superblock": true, 00:34:32.575 "num_base_bdevs": 4, 00:34:32.575 "num_base_bdevs_discovered": 3, 00:34:32.575 "num_base_bdevs_operational": 3, 00:34:32.575 "base_bdevs_list": [ 00:34:32.575 { 00:34:32.575 "name": null, 00:34:32.575 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:32.575 "is_configured": false, 00:34:32.575 "data_offset": 0, 00:34:32.575 "data_size": 63488 00:34:32.575 }, 00:34:32.575 { 00:34:32.575 "name": "BaseBdev2", 00:34:32.575 "uuid": "7558fee5-085f-5ac3-8435-b86409a231aa", 00:34:32.575 "is_configured": true, 00:34:32.575 "data_offset": 2048, 00:34:32.575 "data_size": 63488 00:34:32.575 }, 00:34:32.575 { 00:34:32.575 "name": "BaseBdev3", 00:34:32.575 "uuid": "bfa0591c-6a36-55d2-9f69-00da84d83c39", 00:34:32.575 "is_configured": true, 00:34:32.575 "data_offset": 2048, 00:34:32.575 "data_size": 63488 00:34:32.575 }, 00:34:32.575 { 00:34:32.575 "name": "BaseBdev4", 00:34:32.575 "uuid": "bcaa86a7-b5f3-551b-a89f-783c08db00d0", 00:34:32.575 "is_configured": true, 00:34:32.575 "data_offset": 2048, 00:34:32.575 "data_size": 63488 00:34:32.575 } 00:34:32.575 ] 00:34:32.575 }' 00:34:32.575 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:32.576 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:34:32.576 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:32.576 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:34:32.576 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 82619 00:34:32.576 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 82619 ']' 00:34:32.576 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 82619 00:34:32.576 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:34:32.576 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:34:32.576 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 82619 00:34:32.576 killing process with pid 82619 00:34:32.576 Received shutdown signal, test time was about 60.000000 seconds 00:34:32.576 00:34:32.576 Latency(us) 00:34:32.576 [2024-11-20T13:54:39.635Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:34:32.576 [2024-11-20T13:54:39.635Z] =================================================================================================================== 00:34:32.576 [2024-11-20T13:54:39.635Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:34:32.576 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:34:32.576 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:34:32.576 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 82619' 00:34:32.576 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 82619 00:34:32.576 13:54:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 82619 00:34:32.576 [2024-11-20 13:54:39.631708] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:34:32.576 [2024-11-20 13:54:39.631824] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:34:32.835 [2024-11-20 13:54:39.631895] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:34:32.835 [2024-11-20 13:54:39.631947] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:34:32.835 [2024-11-20 13:54:39.875650] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:34:33.773 13:54:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:34:33.773 00:34:33.773 real 0m24.423s 00:34:33.773 user 0m29.515s 00:34:33.773 sys 0m2.197s 00:34:33.773 13:54:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:34:33.773 13:54:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:34:33.773 ************************************ 00:34:33.773 END TEST raid5f_rebuild_test_sb 00:34:33.773 ************************************ 00:34:33.773 13:54:40 bdev_raid -- bdev/bdev_raid.sh@995 -- # base_blocklen=4096 00:34:33.773 13:54:40 bdev_raid -- bdev/bdev_raid.sh@997 -- # run_test raid_state_function_test_sb_4k raid_state_function_test raid1 2 true 00:34:33.773 13:54:40 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:34:33.773 13:54:40 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:34:33.773 13:54:40 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:34:33.773 ************************************ 00:34:33.773 START TEST raid_state_function_test_sb_4k 00:34:33.773 ************************************ 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:34:33.773 Process raid pid: 83419 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # local strip_size 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@229 -- # raid_pid=83419 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 83419' 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@231 -- # waitforlisten 83419 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@835 -- # '[' -z 83419 ']' 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@840 -- # local max_retries=100 00:34:33.773 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@844 -- # xtrace_disable 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:33.773 13:54:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:34:33.773 [2024-11-20 13:54:40.582606] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:34:33.773 [2024-11-20 13:54:40.582877] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:34:33.774 [2024-11-20 13:54:40.739418] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:34:34.033 [2024-11-20 13:54:40.838476] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:34:34.033 [2024-11-20 13:54:40.961041] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:34:34.033 [2024-11-20 13:54:40.961076] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:34:34.600 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:34:34.600 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@868 -- # return 0 00:34:34.600 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:34:34.600 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:34.600 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:34.600 [2024-11-20 13:54:41.422643] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:34:34.600 [2024-11-20 13:54:41.422701] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:34:34.600 [2024-11-20 13:54:41.422714] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:34:34.600 [2024-11-20 13:54:41.422722] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:34:34.600 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:34.600 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:34:34.600 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:34:34.600 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:34:34.600 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:34:34.600 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:34:34.600 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:34:34.600 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:34.600 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:34.600 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:34.600 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:34.600 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:34.600 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:34:34.600 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:34.600 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:34.600 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:34.600 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:34.600 "name": "Existed_Raid", 00:34:34.600 "uuid": "240f212b-7774-42e1-a1fb-b0fb8426f320", 00:34:34.600 "strip_size_kb": 0, 00:34:34.600 "state": "configuring", 00:34:34.600 "raid_level": "raid1", 00:34:34.600 "superblock": true, 00:34:34.600 "num_base_bdevs": 2, 00:34:34.600 "num_base_bdevs_discovered": 0, 00:34:34.600 "num_base_bdevs_operational": 2, 00:34:34.600 "base_bdevs_list": [ 00:34:34.600 { 00:34:34.600 "name": "BaseBdev1", 00:34:34.600 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:34.600 "is_configured": false, 00:34:34.600 "data_offset": 0, 00:34:34.600 "data_size": 0 00:34:34.600 }, 00:34:34.600 { 00:34:34.600 "name": "BaseBdev2", 00:34:34.600 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:34.600 "is_configured": false, 00:34:34.600 "data_offset": 0, 00:34:34.600 "data_size": 0 00:34:34.600 } 00:34:34.600 ] 00:34:34.600 }' 00:34:34.600 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:34.600 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:34.859 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:34:34.859 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:34.859 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:34.859 [2024-11-20 13:54:41.750650] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:34:34.859 [2024-11-20 13:54:41.750687] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:34:34.859 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:34.859 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:34:34.859 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:34.859 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:34.859 [2024-11-20 13:54:41.758645] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:34:34.859 [2024-11-20 13:54:41.758758] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:34:34.859 [2024-11-20 13:54:41.758811] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:34:34.859 [2024-11-20 13:54:41.758836] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:34:34.859 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:34.859 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev1 00:34:34.859 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:34.859 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:34.859 [2024-11-20 13:54:41.788505] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:34:34.859 BaseBdev1 00:34:34.859 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:34.859 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:34:34.859 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:34:34.859 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:34:34.859 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@905 -- # local i 00:34:34.859 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:34:34.859 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:34:34.859 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:34:34.859 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:34.859 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:34.859 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:34.859 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:34:34.859 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:34.859 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:34.859 [ 00:34:34.859 { 00:34:34.859 "name": "BaseBdev1", 00:34:34.859 "aliases": [ 00:34:34.859 "da90079c-db5d-48c8-a589-401d078f377c" 00:34:34.859 ], 00:34:34.859 "product_name": "Malloc disk", 00:34:34.859 "block_size": 4096, 00:34:34.859 "num_blocks": 8192, 00:34:34.859 "uuid": "da90079c-db5d-48c8-a589-401d078f377c", 00:34:34.859 "assigned_rate_limits": { 00:34:34.859 "rw_ios_per_sec": 0, 00:34:34.859 "rw_mbytes_per_sec": 0, 00:34:34.859 "r_mbytes_per_sec": 0, 00:34:34.859 "w_mbytes_per_sec": 0 00:34:34.859 }, 00:34:34.859 "claimed": true, 00:34:34.859 "claim_type": "exclusive_write", 00:34:34.859 "zoned": false, 00:34:34.859 "supported_io_types": { 00:34:34.859 "read": true, 00:34:34.859 "write": true, 00:34:34.859 "unmap": true, 00:34:34.859 "flush": true, 00:34:34.859 "reset": true, 00:34:34.859 "nvme_admin": false, 00:34:34.859 "nvme_io": false, 00:34:34.859 "nvme_io_md": false, 00:34:34.859 "write_zeroes": true, 00:34:34.859 "zcopy": true, 00:34:34.859 "get_zone_info": false, 00:34:34.859 "zone_management": false, 00:34:34.859 "zone_append": false, 00:34:34.859 "compare": false, 00:34:34.859 "compare_and_write": false, 00:34:34.859 "abort": true, 00:34:34.859 "seek_hole": false, 00:34:34.859 "seek_data": false, 00:34:34.859 "copy": true, 00:34:34.859 "nvme_iov_md": false 00:34:34.859 }, 00:34:34.859 "memory_domains": [ 00:34:34.859 { 00:34:34.859 "dma_device_id": "system", 00:34:34.859 "dma_device_type": 1 00:34:34.860 }, 00:34:34.860 { 00:34:34.860 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:34:34.860 "dma_device_type": 2 00:34:34.860 } 00:34:34.860 ], 00:34:34.860 "driver_specific": {} 00:34:34.860 } 00:34:34.860 ] 00:34:34.860 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:34.860 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@911 -- # return 0 00:34:34.860 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:34:34.860 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:34:34.860 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:34:34.860 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:34:34.860 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:34:34.860 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:34:34.860 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:34.860 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:34.860 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:34.860 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:34.860 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:34.860 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:34:34.860 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:34.860 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:34.860 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:34.860 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:34.860 "name": "Existed_Raid", 00:34:34.860 "uuid": "14ff21f3-510c-4bf6-8249-fe59f58a93d8", 00:34:34.860 "strip_size_kb": 0, 00:34:34.860 "state": "configuring", 00:34:34.860 "raid_level": "raid1", 00:34:34.860 "superblock": true, 00:34:34.860 "num_base_bdevs": 2, 00:34:34.860 "num_base_bdevs_discovered": 1, 00:34:34.860 "num_base_bdevs_operational": 2, 00:34:34.860 "base_bdevs_list": [ 00:34:34.860 { 00:34:34.860 "name": "BaseBdev1", 00:34:34.860 "uuid": "da90079c-db5d-48c8-a589-401d078f377c", 00:34:34.860 "is_configured": true, 00:34:34.860 "data_offset": 256, 00:34:34.860 "data_size": 7936 00:34:34.860 }, 00:34:34.860 { 00:34:34.860 "name": "BaseBdev2", 00:34:34.860 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:34.860 "is_configured": false, 00:34:34.860 "data_offset": 0, 00:34:34.860 "data_size": 0 00:34:34.860 } 00:34:34.860 ] 00:34:34.860 }' 00:34:34.860 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:34.860 13:54:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:35.118 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:34:35.118 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:35.118 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:35.118 [2024-11-20 13:54:42.136613] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:34:35.118 [2024-11-20 13:54:42.136677] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:34:35.118 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:35.118 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:34:35.118 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:35.118 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:35.118 [2024-11-20 13:54:42.144652] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:34:35.118 [2024-11-20 13:54:42.146414] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:34:35.118 [2024-11-20 13:54:42.146450] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:34:35.118 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:35.118 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:34:35.118 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:34:35.118 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:34:35.118 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:34:35.118 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:34:35.118 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:34:35.118 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:34:35.118 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:34:35.118 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:35.118 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:35.118 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:35.118 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:35.118 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:35.118 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:35.118 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:35.118 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:34:35.118 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:35.377 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:35.377 "name": "Existed_Raid", 00:34:35.377 "uuid": "78ee934b-03c0-4159-8b20-723bf9e3c2e0", 00:34:35.377 "strip_size_kb": 0, 00:34:35.377 "state": "configuring", 00:34:35.377 "raid_level": "raid1", 00:34:35.377 "superblock": true, 00:34:35.377 "num_base_bdevs": 2, 00:34:35.377 "num_base_bdevs_discovered": 1, 00:34:35.377 "num_base_bdevs_operational": 2, 00:34:35.377 "base_bdevs_list": [ 00:34:35.377 { 00:34:35.377 "name": "BaseBdev1", 00:34:35.377 "uuid": "da90079c-db5d-48c8-a589-401d078f377c", 00:34:35.377 "is_configured": true, 00:34:35.377 "data_offset": 256, 00:34:35.377 "data_size": 7936 00:34:35.377 }, 00:34:35.377 { 00:34:35.377 "name": "BaseBdev2", 00:34:35.377 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:35.377 "is_configured": false, 00:34:35.377 "data_offset": 0, 00:34:35.377 "data_size": 0 00:34:35.377 } 00:34:35.377 ] 00:34:35.377 }' 00:34:35.377 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:35.377 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev2 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:35.636 [2024-11-20 13:54:42.497141] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:34:35.636 [2024-11-20 13:54:42.497349] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:34:35.636 [2024-11-20 13:54:42.497361] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:34:35.636 [2024-11-20 13:54:42.497588] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:34:35.636 BaseBdev2 00:34:35.636 [2024-11-20 13:54:42.497714] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:34:35.636 [2024-11-20 13:54:42.497725] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:34:35.636 [2024-11-20 13:54:42.497841] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@905 -- # local i 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:35.636 [ 00:34:35.636 { 00:34:35.636 "name": "BaseBdev2", 00:34:35.636 "aliases": [ 00:34:35.636 "40410776-a109-441d-a2b2-a90c47cc526c" 00:34:35.636 ], 00:34:35.636 "product_name": "Malloc disk", 00:34:35.636 "block_size": 4096, 00:34:35.636 "num_blocks": 8192, 00:34:35.636 "uuid": "40410776-a109-441d-a2b2-a90c47cc526c", 00:34:35.636 "assigned_rate_limits": { 00:34:35.636 "rw_ios_per_sec": 0, 00:34:35.636 "rw_mbytes_per_sec": 0, 00:34:35.636 "r_mbytes_per_sec": 0, 00:34:35.636 "w_mbytes_per_sec": 0 00:34:35.636 }, 00:34:35.636 "claimed": true, 00:34:35.636 "claim_type": "exclusive_write", 00:34:35.636 "zoned": false, 00:34:35.636 "supported_io_types": { 00:34:35.636 "read": true, 00:34:35.636 "write": true, 00:34:35.636 "unmap": true, 00:34:35.636 "flush": true, 00:34:35.636 "reset": true, 00:34:35.636 "nvme_admin": false, 00:34:35.636 "nvme_io": false, 00:34:35.636 "nvme_io_md": false, 00:34:35.636 "write_zeroes": true, 00:34:35.636 "zcopy": true, 00:34:35.636 "get_zone_info": false, 00:34:35.636 "zone_management": false, 00:34:35.636 "zone_append": false, 00:34:35.636 "compare": false, 00:34:35.636 "compare_and_write": false, 00:34:35.636 "abort": true, 00:34:35.636 "seek_hole": false, 00:34:35.636 "seek_data": false, 00:34:35.636 "copy": true, 00:34:35.636 "nvme_iov_md": false 00:34:35.636 }, 00:34:35.636 "memory_domains": [ 00:34:35.636 { 00:34:35.636 "dma_device_id": "system", 00:34:35.636 "dma_device_type": 1 00:34:35.636 }, 00:34:35.636 { 00:34:35.636 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:34:35.636 "dma_device_type": 2 00:34:35.636 } 00:34:35.636 ], 00:34:35.636 "driver_specific": {} 00:34:35.636 } 00:34:35.636 ] 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@911 -- # return 0 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:35.636 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:35.636 "name": "Existed_Raid", 00:34:35.636 "uuid": "78ee934b-03c0-4159-8b20-723bf9e3c2e0", 00:34:35.636 "strip_size_kb": 0, 00:34:35.636 "state": "online", 00:34:35.636 "raid_level": "raid1", 00:34:35.636 "superblock": true, 00:34:35.636 "num_base_bdevs": 2, 00:34:35.636 "num_base_bdevs_discovered": 2, 00:34:35.636 "num_base_bdevs_operational": 2, 00:34:35.636 "base_bdevs_list": [ 00:34:35.636 { 00:34:35.636 "name": "BaseBdev1", 00:34:35.637 "uuid": "da90079c-db5d-48c8-a589-401d078f377c", 00:34:35.637 "is_configured": true, 00:34:35.637 "data_offset": 256, 00:34:35.637 "data_size": 7936 00:34:35.637 }, 00:34:35.637 { 00:34:35.637 "name": "BaseBdev2", 00:34:35.637 "uuid": "40410776-a109-441d-a2b2-a90c47cc526c", 00:34:35.637 "is_configured": true, 00:34:35.637 "data_offset": 256, 00:34:35.637 "data_size": 7936 00:34:35.637 } 00:34:35.637 ] 00:34:35.637 }' 00:34:35.637 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:35.637 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@184 -- # local name 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:35.895 [2024-11-20 13:54:42.809474] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:34:35.895 "name": "Existed_Raid", 00:34:35.895 "aliases": [ 00:34:35.895 "78ee934b-03c0-4159-8b20-723bf9e3c2e0" 00:34:35.895 ], 00:34:35.895 "product_name": "Raid Volume", 00:34:35.895 "block_size": 4096, 00:34:35.895 "num_blocks": 7936, 00:34:35.895 "uuid": "78ee934b-03c0-4159-8b20-723bf9e3c2e0", 00:34:35.895 "assigned_rate_limits": { 00:34:35.895 "rw_ios_per_sec": 0, 00:34:35.895 "rw_mbytes_per_sec": 0, 00:34:35.895 "r_mbytes_per_sec": 0, 00:34:35.895 "w_mbytes_per_sec": 0 00:34:35.895 }, 00:34:35.895 "claimed": false, 00:34:35.895 "zoned": false, 00:34:35.895 "supported_io_types": { 00:34:35.895 "read": true, 00:34:35.895 "write": true, 00:34:35.895 "unmap": false, 00:34:35.895 "flush": false, 00:34:35.895 "reset": true, 00:34:35.895 "nvme_admin": false, 00:34:35.895 "nvme_io": false, 00:34:35.895 "nvme_io_md": false, 00:34:35.895 "write_zeroes": true, 00:34:35.895 "zcopy": false, 00:34:35.895 "get_zone_info": false, 00:34:35.895 "zone_management": false, 00:34:35.895 "zone_append": false, 00:34:35.895 "compare": false, 00:34:35.895 "compare_and_write": false, 00:34:35.895 "abort": false, 00:34:35.895 "seek_hole": false, 00:34:35.895 "seek_data": false, 00:34:35.895 "copy": false, 00:34:35.895 "nvme_iov_md": false 00:34:35.895 }, 00:34:35.895 "memory_domains": [ 00:34:35.895 { 00:34:35.895 "dma_device_id": "system", 00:34:35.895 "dma_device_type": 1 00:34:35.895 }, 00:34:35.895 { 00:34:35.895 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:34:35.895 "dma_device_type": 2 00:34:35.895 }, 00:34:35.895 { 00:34:35.895 "dma_device_id": "system", 00:34:35.895 "dma_device_type": 1 00:34:35.895 }, 00:34:35.895 { 00:34:35.895 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:34:35.895 "dma_device_type": 2 00:34:35.895 } 00:34:35.895 ], 00:34:35.895 "driver_specific": { 00:34:35.895 "raid": { 00:34:35.895 "uuid": "78ee934b-03c0-4159-8b20-723bf9e3c2e0", 00:34:35.895 "strip_size_kb": 0, 00:34:35.895 "state": "online", 00:34:35.895 "raid_level": "raid1", 00:34:35.895 "superblock": true, 00:34:35.895 "num_base_bdevs": 2, 00:34:35.895 "num_base_bdevs_discovered": 2, 00:34:35.895 "num_base_bdevs_operational": 2, 00:34:35.895 "base_bdevs_list": [ 00:34:35.895 { 00:34:35.895 "name": "BaseBdev1", 00:34:35.895 "uuid": "da90079c-db5d-48c8-a589-401d078f377c", 00:34:35.895 "is_configured": true, 00:34:35.895 "data_offset": 256, 00:34:35.895 "data_size": 7936 00:34:35.895 }, 00:34:35.895 { 00:34:35.895 "name": "BaseBdev2", 00:34:35.895 "uuid": "40410776-a109-441d-a2b2-a90c47cc526c", 00:34:35.895 "is_configured": true, 00:34:35.895 "data_offset": 256, 00:34:35.895 "data_size": 7936 00:34:35.895 } 00:34:35.895 ] 00:34:35.895 } 00:34:35.895 } 00:34:35.895 }' 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:34:35.895 BaseBdev2' 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:34:35.895 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:36.153 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:34:36.153 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:34:36.153 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:34:36.153 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:36.153 13:54:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:36.153 [2024-11-20 13:54:42.973293] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:34:36.153 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:36.153 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@260 -- # local expected_state 00:34:36.153 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:34:36.153 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@198 -- # case $1 in 00:34:36.153 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@199 -- # return 0 00:34:36.153 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:34:36.153 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:34:36.153 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:34:36.153 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:36.153 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:34:36.153 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:34:36.153 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:34:36.153 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:36.154 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:36.154 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:36.154 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:36.154 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:36.154 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:36.154 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:36.154 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:34:36.154 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:36.154 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:36.154 "name": "Existed_Raid", 00:34:36.154 "uuid": "78ee934b-03c0-4159-8b20-723bf9e3c2e0", 00:34:36.154 "strip_size_kb": 0, 00:34:36.154 "state": "online", 00:34:36.154 "raid_level": "raid1", 00:34:36.154 "superblock": true, 00:34:36.154 "num_base_bdevs": 2, 00:34:36.154 "num_base_bdevs_discovered": 1, 00:34:36.154 "num_base_bdevs_operational": 1, 00:34:36.154 "base_bdevs_list": [ 00:34:36.154 { 00:34:36.154 "name": null, 00:34:36.154 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:36.154 "is_configured": false, 00:34:36.154 "data_offset": 0, 00:34:36.154 "data_size": 7936 00:34:36.154 }, 00:34:36.154 { 00:34:36.154 "name": "BaseBdev2", 00:34:36.154 "uuid": "40410776-a109-441d-a2b2-a90c47cc526c", 00:34:36.154 "is_configured": true, 00:34:36.154 "data_offset": 256, 00:34:36.154 "data_size": 7936 00:34:36.154 } 00:34:36.154 ] 00:34:36.154 }' 00:34:36.154 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:36.154 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:36.412 [2024-11-20 13:54:43.358762] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:34:36.412 [2024-11-20 13:54:43.358964] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:34:36.412 [2024-11-20 13:54:43.407490] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:34:36.412 [2024-11-20 13:54:43.407615] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:34:36.412 [2024-11-20 13:54:43.407688] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@326 -- # killprocess 83419 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@954 -- # '[' -z 83419 ']' 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@958 -- # kill -0 83419 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@959 -- # uname 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:34:36.412 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 83419 00:34:36.670 killing process with pid 83419 00:34:36.670 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:34:36.670 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:34:36.670 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@972 -- # echo 'killing process with pid 83419' 00:34:36.670 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@973 -- # kill 83419 00:34:36.670 [2024-11-20 13:54:43.477610] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:34:36.670 13:54:43 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@978 -- # wait 83419 00:34:36.670 [2024-11-20 13:54:43.486495] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:34:37.236 13:54:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@328 -- # return 0 00:34:37.236 00:34:37.236 real 0m3.578s 00:34:37.236 user 0m5.169s 00:34:37.236 sys 0m0.630s 00:34:37.236 13:54:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@1130 -- # xtrace_disable 00:34:37.236 ************************************ 00:34:37.236 END TEST raid_state_function_test_sb_4k 00:34:37.236 ************************************ 00:34:37.236 13:54:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:37.236 13:54:44 bdev_raid -- bdev/bdev_raid.sh@998 -- # run_test raid_superblock_test_4k raid_superblock_test raid1 2 00:34:37.236 13:54:44 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:34:37.236 13:54:44 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:34:37.236 13:54:44 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:34:37.236 ************************************ 00:34:37.236 START TEST raid_superblock_test_4k 00:34:37.236 ************************************ 00:34:37.236 13:54:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:34:37.236 13:54:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:34:37.236 13:54:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:34:37.236 13:54:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:34:37.236 13:54:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:34:37.236 13:54:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:34:37.236 13:54:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:34:37.236 13:54:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:34:37.236 13:54:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:34:37.236 13:54:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:34:37.236 13:54:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@399 -- # local strip_size 00:34:37.236 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:34:37.236 13:54:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:34:37.236 13:54:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:34:37.236 13:54:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:34:37.236 13:54:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:34:37.236 13:54:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:34:37.236 13:54:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@412 -- # raid_pid=83649 00:34:37.236 13:54:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@413 -- # waitforlisten 83649 00:34:37.236 13:54:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@835 -- # '[' -z 83649 ']' 00:34:37.236 13:54:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:34:37.236 13:54:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@840 -- # local max_retries=100 00:34:37.236 13:54:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:34:37.236 13:54:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@844 -- # xtrace_disable 00:34:37.236 13:54:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:37.236 13:54:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:34:37.236 [2024-11-20 13:54:44.200404] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:34:37.236 [2024-11-20 13:54:44.200535] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid83649 ] 00:34:37.494 [2024-11-20 13:54:44.358148] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:34:37.494 [2024-11-20 13:54:44.458549] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:34:37.752 [2024-11-20 13:54:44.577457] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:34:37.752 [2024-11-20 13:54:44.577670] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:34:38.010 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:34:38.010 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@868 -- # return 0 00:34:38.010 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:34:38.010 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:34:38.010 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:34:38.010 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:34:38.010 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:34:38.010 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:34:38.010 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:34:38.010 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:34:38.010 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc1 00:34:38.010 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:38.010 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:38.010 malloc1 00:34:38.010 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:38.010 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:34:38.010 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:38.010 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:38.268 [2024-11-20 13:54:45.070645] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:34:38.269 [2024-11-20 13:54:45.070708] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:34:38.269 [2024-11-20 13:54:45.070729] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:34:38.269 [2024-11-20 13:54:45.070737] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:34:38.269 [2024-11-20 13:54:45.072629] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:34:38.269 [2024-11-20 13:54:45.072666] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:34:38.269 pt1 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc2 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:38.269 malloc2 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:38.269 [2024-11-20 13:54:45.107952] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:34:38.269 [2024-11-20 13:54:45.108011] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:34:38.269 [2024-11-20 13:54:45.108035] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:34:38.269 [2024-11-20 13:54:45.108042] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:34:38.269 [2024-11-20 13:54:45.109799] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:34:38.269 [2024-11-20 13:54:45.109826] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:34:38.269 pt2 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:38.269 [2024-11-20 13:54:45.116009] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:34:38.269 [2024-11-20 13:54:45.117566] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:34:38.269 [2024-11-20 13:54:45.117693] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:34:38.269 [2024-11-20 13:54:45.117706] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:34:38.269 [2024-11-20 13:54:45.117906] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:34:38.269 [2024-11-20 13:54:45.118188] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:34:38.269 [2024-11-20 13:54:45.118509] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:34:38.269 [2024-11-20 13:54:45.118729] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:38.269 "name": "raid_bdev1", 00:34:38.269 "uuid": "c3d2dcff-95f4-4bd4-887d-0b08275ad687", 00:34:38.269 "strip_size_kb": 0, 00:34:38.269 "state": "online", 00:34:38.269 "raid_level": "raid1", 00:34:38.269 "superblock": true, 00:34:38.269 "num_base_bdevs": 2, 00:34:38.269 "num_base_bdevs_discovered": 2, 00:34:38.269 "num_base_bdevs_operational": 2, 00:34:38.269 "base_bdevs_list": [ 00:34:38.269 { 00:34:38.269 "name": "pt1", 00:34:38.269 "uuid": "00000000-0000-0000-0000-000000000001", 00:34:38.269 "is_configured": true, 00:34:38.269 "data_offset": 256, 00:34:38.269 "data_size": 7936 00:34:38.269 }, 00:34:38.269 { 00:34:38.269 "name": "pt2", 00:34:38.269 "uuid": "00000000-0000-0000-0000-000000000002", 00:34:38.269 "is_configured": true, 00:34:38.269 "data_offset": 256, 00:34:38.269 "data_size": 7936 00:34:38.269 } 00:34:38.269 ] 00:34:38.269 }' 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:38.269 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@184 -- # local name 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:38.628 [2024-11-20 13:54:45.460340] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:34:38.628 "name": "raid_bdev1", 00:34:38.628 "aliases": [ 00:34:38.628 "c3d2dcff-95f4-4bd4-887d-0b08275ad687" 00:34:38.628 ], 00:34:38.628 "product_name": "Raid Volume", 00:34:38.628 "block_size": 4096, 00:34:38.628 "num_blocks": 7936, 00:34:38.628 "uuid": "c3d2dcff-95f4-4bd4-887d-0b08275ad687", 00:34:38.628 "assigned_rate_limits": { 00:34:38.628 "rw_ios_per_sec": 0, 00:34:38.628 "rw_mbytes_per_sec": 0, 00:34:38.628 "r_mbytes_per_sec": 0, 00:34:38.628 "w_mbytes_per_sec": 0 00:34:38.628 }, 00:34:38.628 "claimed": false, 00:34:38.628 "zoned": false, 00:34:38.628 "supported_io_types": { 00:34:38.628 "read": true, 00:34:38.628 "write": true, 00:34:38.628 "unmap": false, 00:34:38.628 "flush": false, 00:34:38.628 "reset": true, 00:34:38.628 "nvme_admin": false, 00:34:38.628 "nvme_io": false, 00:34:38.628 "nvme_io_md": false, 00:34:38.628 "write_zeroes": true, 00:34:38.628 "zcopy": false, 00:34:38.628 "get_zone_info": false, 00:34:38.628 "zone_management": false, 00:34:38.628 "zone_append": false, 00:34:38.628 "compare": false, 00:34:38.628 "compare_and_write": false, 00:34:38.628 "abort": false, 00:34:38.628 "seek_hole": false, 00:34:38.628 "seek_data": false, 00:34:38.628 "copy": false, 00:34:38.628 "nvme_iov_md": false 00:34:38.628 }, 00:34:38.628 "memory_domains": [ 00:34:38.628 { 00:34:38.628 "dma_device_id": "system", 00:34:38.628 "dma_device_type": 1 00:34:38.628 }, 00:34:38.628 { 00:34:38.628 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:34:38.628 "dma_device_type": 2 00:34:38.628 }, 00:34:38.628 { 00:34:38.628 "dma_device_id": "system", 00:34:38.628 "dma_device_type": 1 00:34:38.628 }, 00:34:38.628 { 00:34:38.628 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:34:38.628 "dma_device_type": 2 00:34:38.628 } 00:34:38.628 ], 00:34:38.628 "driver_specific": { 00:34:38.628 "raid": { 00:34:38.628 "uuid": "c3d2dcff-95f4-4bd4-887d-0b08275ad687", 00:34:38.628 "strip_size_kb": 0, 00:34:38.628 "state": "online", 00:34:38.628 "raid_level": "raid1", 00:34:38.628 "superblock": true, 00:34:38.628 "num_base_bdevs": 2, 00:34:38.628 "num_base_bdevs_discovered": 2, 00:34:38.628 "num_base_bdevs_operational": 2, 00:34:38.628 "base_bdevs_list": [ 00:34:38.628 { 00:34:38.628 "name": "pt1", 00:34:38.628 "uuid": "00000000-0000-0000-0000-000000000001", 00:34:38.628 "is_configured": true, 00:34:38.628 "data_offset": 256, 00:34:38.628 "data_size": 7936 00:34:38.628 }, 00:34:38.628 { 00:34:38.628 "name": "pt2", 00:34:38.628 "uuid": "00000000-0000-0000-0000-000000000002", 00:34:38.628 "is_configured": true, 00:34:38.628 "data_offset": 256, 00:34:38.628 "data_size": 7936 00:34:38.628 } 00:34:38.628 ] 00:34:38.628 } 00:34:38.628 } 00:34:38.628 }' 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:34:38.628 pt2' 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:38.628 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:38.628 [2024-11-20 13:54:45.628288] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:34:38.629 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:38.629 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=c3d2dcff-95f4-4bd4-887d-0b08275ad687 00:34:38.629 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@436 -- # '[' -z c3d2dcff-95f4-4bd4-887d-0b08275ad687 ']' 00:34:38.629 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:34:38.629 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:38.629 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:38.899 [2024-11-20 13:54:45.656049] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:34:38.899 [2024-11-20 13:54:45.656067] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:34:38.899 [2024-11-20 13:54:45.656138] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:34:38.899 [2024-11-20 13:54:45.656196] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:34:38.899 [2024-11-20 13:54:45.656206] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@652 -- # local es=0 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:38.899 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:38.899 [2024-11-20 13:54:45.752099] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:34:38.899 [2024-11-20 13:54:45.753855] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:34:38.899 [2024-11-20 13:54:45.753914] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:34:38.899 [2024-11-20 13:54:45.753960] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:34:38.899 [2024-11-20 13:54:45.753972] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:34:38.899 [2024-11-20 13:54:45.753991] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:34:38.899 request: 00:34:38.899 { 00:34:38.899 "name": "raid_bdev1", 00:34:38.899 "raid_level": "raid1", 00:34:38.899 "base_bdevs": [ 00:34:38.899 "malloc1", 00:34:38.899 "malloc2" 00:34:38.899 ], 00:34:38.899 "superblock": false, 00:34:38.899 "method": "bdev_raid_create", 00:34:38.899 "req_id": 1 00:34:38.899 } 00:34:38.899 Got JSON-RPC error response 00:34:38.899 response: 00:34:38.899 { 00:34:38.900 "code": -17, 00:34:38.900 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:34:38.900 } 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@655 -- # es=1 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:38.900 [2024-11-20 13:54:45.800095] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:34:38.900 [2024-11-20 13:54:45.800136] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:34:38.900 [2024-11-20 13:54:45.800153] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:34:38.900 [2024-11-20 13:54:45.800162] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:34:38.900 [2024-11-20 13:54:45.802074] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:34:38.900 [2024-11-20 13:54:45.802103] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:34:38.900 [2024-11-20 13:54:45.802165] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:34:38.900 [2024-11-20 13:54:45.802211] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:34:38.900 pt1 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:38.900 "name": "raid_bdev1", 00:34:38.900 "uuid": "c3d2dcff-95f4-4bd4-887d-0b08275ad687", 00:34:38.900 "strip_size_kb": 0, 00:34:38.900 "state": "configuring", 00:34:38.900 "raid_level": "raid1", 00:34:38.900 "superblock": true, 00:34:38.900 "num_base_bdevs": 2, 00:34:38.900 "num_base_bdevs_discovered": 1, 00:34:38.900 "num_base_bdevs_operational": 2, 00:34:38.900 "base_bdevs_list": [ 00:34:38.900 { 00:34:38.900 "name": "pt1", 00:34:38.900 "uuid": "00000000-0000-0000-0000-000000000001", 00:34:38.900 "is_configured": true, 00:34:38.900 "data_offset": 256, 00:34:38.900 "data_size": 7936 00:34:38.900 }, 00:34:38.900 { 00:34:38.900 "name": null, 00:34:38.900 "uuid": "00000000-0000-0000-0000-000000000002", 00:34:38.900 "is_configured": false, 00:34:38.900 "data_offset": 256, 00:34:38.900 "data_size": 7936 00:34:38.900 } 00:34:38.900 ] 00:34:38.900 }' 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:38.900 13:54:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:39.157 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:34:39.157 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:34:39.157 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:34:39.157 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:34:39.157 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:39.157 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:39.157 [2024-11-20 13:54:46.088168] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:34:39.157 [2024-11-20 13:54:46.088226] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:34:39.157 [2024-11-20 13:54:46.088245] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:34:39.157 [2024-11-20 13:54:46.088255] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:34:39.157 [2024-11-20 13:54:46.088644] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:34:39.157 [2024-11-20 13:54:46.088659] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:34:39.157 [2024-11-20 13:54:46.088724] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:34:39.157 [2024-11-20 13:54:46.088746] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:34:39.157 [2024-11-20 13:54:46.088843] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:34:39.157 [2024-11-20 13:54:46.088852] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:34:39.157 [2024-11-20 13:54:46.089071] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:34:39.157 [2024-11-20 13:54:46.089186] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:34:39.157 [2024-11-20 13:54:46.089194] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:34:39.157 [2024-11-20 13:54:46.089303] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:34:39.157 pt2 00:34:39.157 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:39.157 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:34:39.158 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:34:39.158 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:34:39.158 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:39.158 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:39.158 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:34:39.158 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:34:39.158 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:34:39.158 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:39.158 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:39.158 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:39.158 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:39.158 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:39.158 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:39.158 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:39.158 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:39.158 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:39.158 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:39.158 "name": "raid_bdev1", 00:34:39.158 "uuid": "c3d2dcff-95f4-4bd4-887d-0b08275ad687", 00:34:39.158 "strip_size_kb": 0, 00:34:39.158 "state": "online", 00:34:39.158 "raid_level": "raid1", 00:34:39.158 "superblock": true, 00:34:39.158 "num_base_bdevs": 2, 00:34:39.158 "num_base_bdevs_discovered": 2, 00:34:39.158 "num_base_bdevs_operational": 2, 00:34:39.158 "base_bdevs_list": [ 00:34:39.158 { 00:34:39.158 "name": "pt1", 00:34:39.158 "uuid": "00000000-0000-0000-0000-000000000001", 00:34:39.158 "is_configured": true, 00:34:39.158 "data_offset": 256, 00:34:39.158 "data_size": 7936 00:34:39.158 }, 00:34:39.158 { 00:34:39.158 "name": "pt2", 00:34:39.158 "uuid": "00000000-0000-0000-0000-000000000002", 00:34:39.158 "is_configured": true, 00:34:39.158 "data_offset": 256, 00:34:39.158 "data_size": 7936 00:34:39.158 } 00:34:39.158 ] 00:34:39.158 }' 00:34:39.158 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:39.158 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:39.415 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:34:39.415 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:34:39.415 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:34:39.415 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:34:39.415 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@184 -- # local name 00:34:39.415 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:34:39.415 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:34:39.415 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:34:39.415 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:39.415 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:39.415 [2024-11-20 13:54:46.420471] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:34:39.415 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:39.415 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:34:39.415 "name": "raid_bdev1", 00:34:39.415 "aliases": [ 00:34:39.415 "c3d2dcff-95f4-4bd4-887d-0b08275ad687" 00:34:39.415 ], 00:34:39.415 "product_name": "Raid Volume", 00:34:39.415 "block_size": 4096, 00:34:39.415 "num_blocks": 7936, 00:34:39.415 "uuid": "c3d2dcff-95f4-4bd4-887d-0b08275ad687", 00:34:39.415 "assigned_rate_limits": { 00:34:39.415 "rw_ios_per_sec": 0, 00:34:39.415 "rw_mbytes_per_sec": 0, 00:34:39.415 "r_mbytes_per_sec": 0, 00:34:39.415 "w_mbytes_per_sec": 0 00:34:39.415 }, 00:34:39.415 "claimed": false, 00:34:39.415 "zoned": false, 00:34:39.415 "supported_io_types": { 00:34:39.415 "read": true, 00:34:39.415 "write": true, 00:34:39.415 "unmap": false, 00:34:39.415 "flush": false, 00:34:39.415 "reset": true, 00:34:39.415 "nvme_admin": false, 00:34:39.415 "nvme_io": false, 00:34:39.415 "nvme_io_md": false, 00:34:39.415 "write_zeroes": true, 00:34:39.415 "zcopy": false, 00:34:39.415 "get_zone_info": false, 00:34:39.415 "zone_management": false, 00:34:39.415 "zone_append": false, 00:34:39.415 "compare": false, 00:34:39.415 "compare_and_write": false, 00:34:39.415 "abort": false, 00:34:39.415 "seek_hole": false, 00:34:39.415 "seek_data": false, 00:34:39.415 "copy": false, 00:34:39.415 "nvme_iov_md": false 00:34:39.415 }, 00:34:39.415 "memory_domains": [ 00:34:39.415 { 00:34:39.415 "dma_device_id": "system", 00:34:39.415 "dma_device_type": 1 00:34:39.415 }, 00:34:39.415 { 00:34:39.415 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:34:39.415 "dma_device_type": 2 00:34:39.415 }, 00:34:39.415 { 00:34:39.415 "dma_device_id": "system", 00:34:39.415 "dma_device_type": 1 00:34:39.415 }, 00:34:39.415 { 00:34:39.415 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:34:39.415 "dma_device_type": 2 00:34:39.415 } 00:34:39.415 ], 00:34:39.415 "driver_specific": { 00:34:39.415 "raid": { 00:34:39.415 "uuid": "c3d2dcff-95f4-4bd4-887d-0b08275ad687", 00:34:39.415 "strip_size_kb": 0, 00:34:39.415 "state": "online", 00:34:39.415 "raid_level": "raid1", 00:34:39.415 "superblock": true, 00:34:39.415 "num_base_bdevs": 2, 00:34:39.415 "num_base_bdevs_discovered": 2, 00:34:39.415 "num_base_bdevs_operational": 2, 00:34:39.415 "base_bdevs_list": [ 00:34:39.415 { 00:34:39.415 "name": "pt1", 00:34:39.415 "uuid": "00000000-0000-0000-0000-000000000001", 00:34:39.415 "is_configured": true, 00:34:39.415 "data_offset": 256, 00:34:39.415 "data_size": 7936 00:34:39.415 }, 00:34:39.415 { 00:34:39.415 "name": "pt2", 00:34:39.415 "uuid": "00000000-0000-0000-0000-000000000002", 00:34:39.415 "is_configured": true, 00:34:39.415 "data_offset": 256, 00:34:39.415 "data_size": 7936 00:34:39.415 } 00:34:39.415 ] 00:34:39.415 } 00:34:39.415 } 00:34:39.415 }' 00:34:39.415 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:34:39.673 pt2' 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:39.673 [2024-11-20 13:54:46.584448] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # '[' c3d2dcff-95f4-4bd4-887d-0b08275ad687 '!=' c3d2dcff-95f4-4bd4-887d-0b08275ad687 ']' 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@198 -- # case $1 in 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@199 -- # return 0 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:39.673 [2024-11-20 13:54:46.616284] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:39.673 "name": "raid_bdev1", 00:34:39.673 "uuid": "c3d2dcff-95f4-4bd4-887d-0b08275ad687", 00:34:39.673 "strip_size_kb": 0, 00:34:39.673 "state": "online", 00:34:39.673 "raid_level": "raid1", 00:34:39.673 "superblock": true, 00:34:39.673 "num_base_bdevs": 2, 00:34:39.673 "num_base_bdevs_discovered": 1, 00:34:39.673 "num_base_bdevs_operational": 1, 00:34:39.673 "base_bdevs_list": [ 00:34:39.673 { 00:34:39.673 "name": null, 00:34:39.673 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:39.673 "is_configured": false, 00:34:39.673 "data_offset": 0, 00:34:39.673 "data_size": 7936 00:34:39.673 }, 00:34:39.673 { 00:34:39.673 "name": "pt2", 00:34:39.673 "uuid": "00000000-0000-0000-0000-000000000002", 00:34:39.673 "is_configured": true, 00:34:39.673 "data_offset": 256, 00:34:39.673 "data_size": 7936 00:34:39.673 } 00:34:39.673 ] 00:34:39.673 }' 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:39.673 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:39.931 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:34:39.931 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:39.931 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:39.931 [2024-11-20 13:54:46.952345] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:34:39.931 [2024-11-20 13:54:46.952373] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:34:39.931 [2024-11-20 13:54:46.952444] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:34:39.931 [2024-11-20 13:54:46.952489] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:34:39.931 [2024-11-20 13:54:46.952500] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:34:39.931 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:39.931 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:39.931 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:39.931 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:39.931 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:34:39.931 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:40.189 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:34:40.189 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:34:40.189 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:34:40.189 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:34:40.189 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:34:40.189 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:40.189 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:40.189 13:54:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:40.189 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:34:40.189 13:54:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:34:40.189 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:34:40.189 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:34:40.189 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@519 -- # i=1 00:34:40.189 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:34:40.189 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:40.189 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:40.189 [2024-11-20 13:54:47.004320] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:34:40.189 [2024-11-20 13:54:47.004370] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:34:40.189 [2024-11-20 13:54:47.004386] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:34:40.189 [2024-11-20 13:54:47.004395] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:34:40.189 [2024-11-20 13:54:47.006398] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:34:40.189 [2024-11-20 13:54:47.006432] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:34:40.189 [2024-11-20 13:54:47.006499] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:34:40.189 [2024-11-20 13:54:47.006540] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:34:40.189 [2024-11-20 13:54:47.006624] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:34:40.189 [2024-11-20 13:54:47.006635] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:34:40.189 [2024-11-20 13:54:47.006835] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:34:40.189 [2024-11-20 13:54:47.006947] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:34:40.189 [2024-11-20 13:54:47.006954] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:34:40.189 [2024-11-20 13:54:47.007074] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:34:40.189 pt2 00:34:40.189 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:40.189 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:34:40.189 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:40.189 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:40.190 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:34:40.190 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:34:40.190 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:34:40.190 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:40.190 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:40.190 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:40.190 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:40.190 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:40.190 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:40.190 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:40.190 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:40.190 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:40.190 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:40.190 "name": "raid_bdev1", 00:34:40.190 "uuid": "c3d2dcff-95f4-4bd4-887d-0b08275ad687", 00:34:40.190 "strip_size_kb": 0, 00:34:40.190 "state": "online", 00:34:40.190 "raid_level": "raid1", 00:34:40.190 "superblock": true, 00:34:40.190 "num_base_bdevs": 2, 00:34:40.190 "num_base_bdevs_discovered": 1, 00:34:40.190 "num_base_bdevs_operational": 1, 00:34:40.190 "base_bdevs_list": [ 00:34:40.190 { 00:34:40.190 "name": null, 00:34:40.190 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:40.190 "is_configured": false, 00:34:40.190 "data_offset": 256, 00:34:40.190 "data_size": 7936 00:34:40.190 }, 00:34:40.190 { 00:34:40.190 "name": "pt2", 00:34:40.190 "uuid": "00000000-0000-0000-0000-000000000002", 00:34:40.190 "is_configured": true, 00:34:40.190 "data_offset": 256, 00:34:40.190 "data_size": 7936 00:34:40.190 } 00:34:40.190 ] 00:34:40.190 }' 00:34:40.190 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:40.190 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:40.448 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:34:40.448 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:40.448 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:40.448 [2024-11-20 13:54:47.308382] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:34:40.448 [2024-11-20 13:54:47.308408] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:34:40.448 [2024-11-20 13:54:47.308476] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:34:40.448 [2024-11-20 13:54:47.308522] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:34:40.448 [2024-11-20 13:54:47.308531] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:34:40.448 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:40.449 [2024-11-20 13:54:47.356396] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:34:40.449 [2024-11-20 13:54:47.356533] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:34:40.449 [2024-11-20 13:54:47.356571] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:34:40.449 [2024-11-20 13:54:47.356918] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:34:40.449 [2024-11-20 13:54:47.359043] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:34:40.449 [2024-11-20 13:54:47.359135] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:34:40.449 [2024-11-20 13:54:47.359229] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:34:40.449 [2024-11-20 13:54:47.359273] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:34:40.449 [2024-11-20 13:54:47.359399] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:34:40.449 [2024-11-20 13:54:47.359408] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:34:40.449 [2024-11-20 13:54:47.359423] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:34:40.449 [2024-11-20 13:54:47.359464] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:34:40.449 [2024-11-20 13:54:47.359533] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:34:40.449 [2024-11-20 13:54:47.359540] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:34:40.449 [2024-11-20 13:54:47.359757] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:34:40.449 [2024-11-20 13:54:47.359871] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:34:40.449 [2024-11-20 13:54:47.359880] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:34:40.449 [2024-11-20 13:54:47.360009] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:34:40.449 pt1 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:40.449 "name": "raid_bdev1", 00:34:40.449 "uuid": "c3d2dcff-95f4-4bd4-887d-0b08275ad687", 00:34:40.449 "strip_size_kb": 0, 00:34:40.449 "state": "online", 00:34:40.449 "raid_level": "raid1", 00:34:40.449 "superblock": true, 00:34:40.449 "num_base_bdevs": 2, 00:34:40.449 "num_base_bdevs_discovered": 1, 00:34:40.449 "num_base_bdevs_operational": 1, 00:34:40.449 "base_bdevs_list": [ 00:34:40.449 { 00:34:40.449 "name": null, 00:34:40.449 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:40.449 "is_configured": false, 00:34:40.449 "data_offset": 256, 00:34:40.449 "data_size": 7936 00:34:40.449 }, 00:34:40.449 { 00:34:40.449 "name": "pt2", 00:34:40.449 "uuid": "00000000-0000-0000-0000-000000000002", 00:34:40.449 "is_configured": true, 00:34:40.449 "data_offset": 256, 00:34:40.449 "data_size": 7936 00:34:40.449 } 00:34:40.449 ] 00:34:40.449 }' 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:40.449 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:40.708 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:34:40.708 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:34:40.708 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:40.708 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:40.708 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:40.708 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:34:40.708 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:34:40.708 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:34:40.708 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:40.708 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:40.708 [2024-11-20 13:54:47.700674] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:34:40.708 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:40.708 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # '[' c3d2dcff-95f4-4bd4-887d-0b08275ad687 '!=' c3d2dcff-95f4-4bd4-887d-0b08275ad687 ']' 00:34:40.708 13:54:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@563 -- # killprocess 83649 00:34:40.708 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@954 -- # '[' -z 83649 ']' 00:34:40.708 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@958 -- # kill -0 83649 00:34:40.708 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@959 -- # uname 00:34:40.708 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:34:40.708 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 83649 00:34:40.708 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:34:40.708 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:34:40.708 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@972 -- # echo 'killing process with pid 83649' 00:34:40.708 killing process with pid 83649 00:34:40.708 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@973 -- # kill 83649 00:34:40.708 [2024-11-20 13:54:47.753364] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:34:40.708 13:54:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@978 -- # wait 83649 00:34:40.708 [2024-11-20 13:54:47.753525] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:34:40.708 [2024-11-20 13:54:47.753574] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:34:40.708 [2024-11-20 13:54:47.753590] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:34:40.966 [2024-11-20 13:54:47.859663] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:34:41.533 ************************************ 00:34:41.533 END TEST raid_superblock_test_4k 00:34:41.533 ************************************ 00:34:41.533 13:54:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@565 -- # return 0 00:34:41.533 00:34:41.533 real 0m4.332s 00:34:41.533 user 0m6.589s 00:34:41.533 sys 0m0.752s 00:34:41.533 13:54:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@1130 -- # xtrace_disable 00:34:41.533 13:54:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:34:41.533 13:54:48 bdev_raid -- bdev/bdev_raid.sh@999 -- # '[' true = true ']' 00:34:41.533 13:54:48 bdev_raid -- bdev/bdev_raid.sh@1000 -- # run_test raid_rebuild_test_sb_4k raid_rebuild_test raid1 2 true false true 00:34:41.533 13:54:48 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:34:41.533 13:54:48 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:34:41.533 13:54:48 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:34:41.533 ************************************ 00:34:41.533 START TEST raid_rebuild_test_sb_4k 00:34:41.533 ************************************ 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false true 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@573 -- # local verify=true 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:34:41.533 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # local strip_size 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@577 -- # local create_arg 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@579 -- # local data_offset 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@597 -- # raid_pid=83961 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@598 -- # waitforlisten 83961 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@835 -- # '[' -z 83961 ']' 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@840 -- # local max_retries=100 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@844 -- # xtrace_disable 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:41.533 13:54:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:34:41.533 [2024-11-20 13:54:48.578595] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:34:41.533 I/O size of 3145728 is greater than zero copy threshold (65536). 00:34:41.533 Zero copy mechanism will not be used. 00:34:41.533 [2024-11-20 13:54:48.578895] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid83961 ] 00:34:41.791 [2024-11-20 13:54:48.735173] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:34:41.791 [2024-11-20 13:54:48.838915] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:34:42.050 [2024-11-20 13:54:48.959201] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:34:42.050 [2024-11-20 13:54:48.959247] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@868 -- # return 0 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev1_malloc 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:42.675 BaseBdev1_malloc 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:42.675 [2024-11-20 13:54:49.462789] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:34:42.675 [2024-11-20 13:54:49.462852] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:34:42.675 [2024-11-20 13:54:49.462872] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:34:42.675 [2024-11-20 13:54:49.462881] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:34:42.675 [2024-11-20 13:54:49.464746] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:34:42.675 [2024-11-20 13:54:49.464778] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:34:42.675 BaseBdev1 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev2_malloc 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:42.675 BaseBdev2_malloc 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:42.675 [2024-11-20 13:54:49.496281] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:34:42.675 [2024-11-20 13:54:49.496331] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:34:42.675 [2024-11-20 13:54:49.496349] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:34:42.675 [2024-11-20 13:54:49.496358] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:34:42.675 [2024-11-20 13:54:49.498214] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:34:42.675 [2024-11-20 13:54:49.498243] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:34:42.675 BaseBdev2 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -b spare_malloc 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:42.675 spare_malloc 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:42.675 spare_delay 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:42.675 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:42.675 [2024-11-20 13:54:49.549642] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:34:42.675 [2024-11-20 13:54:49.549696] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:34:42.675 [2024-11-20 13:54:49.549711] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:34:42.675 [2024-11-20 13:54:49.549721] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:34:42.676 [2024-11-20 13:54:49.551573] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:34:42.676 [2024-11-20 13:54:49.551733] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:34:42.676 spare 00:34:42.676 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:42.676 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:34:42.676 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:42.676 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:42.676 [2024-11-20 13:54:49.557695] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:34:42.676 [2024-11-20 13:54:49.559362] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:34:42.676 [2024-11-20 13:54:49.559566] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:34:42.676 [2024-11-20 13:54:49.559626] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:34:42.676 [2024-11-20 13:54:49.559877] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:34:42.676 [2024-11-20 13:54:49.560078] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:34:42.676 [2024-11-20 13:54:49.560135] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:34:42.676 [2024-11-20 13:54:49.560301] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:34:42.676 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:42.676 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:34:42.676 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:42.676 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:42.676 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:34:42.676 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:34:42.676 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:34:42.676 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:42.676 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:42.676 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:42.676 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:42.676 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:42.676 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:42.676 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:42.676 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:42.676 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:42.676 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:42.676 "name": "raid_bdev1", 00:34:42.676 "uuid": "86cc1961-187e-4be6-b876-3fdac7636128", 00:34:42.676 "strip_size_kb": 0, 00:34:42.676 "state": "online", 00:34:42.676 "raid_level": "raid1", 00:34:42.676 "superblock": true, 00:34:42.676 "num_base_bdevs": 2, 00:34:42.676 "num_base_bdevs_discovered": 2, 00:34:42.676 "num_base_bdevs_operational": 2, 00:34:42.676 "base_bdevs_list": [ 00:34:42.676 { 00:34:42.676 "name": "BaseBdev1", 00:34:42.676 "uuid": "2af07cf7-0f0f-5520-a14f-5bc14f896feb", 00:34:42.676 "is_configured": true, 00:34:42.676 "data_offset": 256, 00:34:42.676 "data_size": 7936 00:34:42.676 }, 00:34:42.676 { 00:34:42.676 "name": "BaseBdev2", 00:34:42.676 "uuid": "462c1e65-ee53-50fd-b426-a46b5461db00", 00:34:42.676 "is_configured": true, 00:34:42.676 "data_offset": 256, 00:34:42.676 "data_size": 7936 00:34:42.676 } 00:34:42.676 ] 00:34:42.676 }' 00:34:42.676 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:42.676 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:42.935 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:34:42.935 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:34:42.935 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:42.935 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:42.935 [2024-11-20 13:54:49.866044] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:34:42.935 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:42.935 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:34:42.935 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:42.935 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:34:42.935 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:42.935 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:42.935 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:42.935 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:34:42.935 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:34:42.935 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:34:42.935 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:34:42.935 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:34:42.935 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:34:42.935 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:34:42.935 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # local bdev_list 00:34:42.935 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:34:42.935 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # local nbd_list 00:34:42.935 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@12 -- # local i 00:34:42.935 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:34:42.935 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:34:42.935 13:54:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:34:43.194 [2024-11-20 13:54:50.073860] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:34:43.194 /dev/nbd0 00:34:43.194 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:34:43.194 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:34:43.194 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:34:43.194 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # local i 00:34:43.194 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:34:43.194 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:34:43.194 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:34:43.194 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@877 -- # break 00:34:43.194 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:34:43.194 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:34:43.194 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:34:43.194 1+0 records in 00:34:43.194 1+0 records out 00:34:43.194 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00017301 s, 23.7 MB/s 00:34:43.194 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:34:43.194 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # size=4096 00:34:43.194 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:34:43.194 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:34:43.194 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@893 -- # return 0 00:34:43.194 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:34:43.194 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:34:43.194 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:34:43.194 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:34:43.194 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=4096 count=7936 oflag=direct 00:34:43.761 7936+0 records in 00:34:43.761 7936+0 records out 00:34:43.761 32505856 bytes (33 MB, 31 MiB) copied, 0.588646 s, 55.2 MB/s 00:34:43.761 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:34:43.761 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:34:43.761 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:34:43.761 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # local nbd_list 00:34:43.761 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@51 -- # local i 00:34:43.761 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:34:43.761 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:34:44.020 [2024-11-20 13:54:50.902685] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:44.020 [2024-11-20 13:54:50.933018] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:44.020 "name": "raid_bdev1", 00:34:44.020 "uuid": "86cc1961-187e-4be6-b876-3fdac7636128", 00:34:44.020 "strip_size_kb": 0, 00:34:44.020 "state": "online", 00:34:44.020 "raid_level": "raid1", 00:34:44.020 "superblock": true, 00:34:44.020 "num_base_bdevs": 2, 00:34:44.020 "num_base_bdevs_discovered": 1, 00:34:44.020 "num_base_bdevs_operational": 1, 00:34:44.020 "base_bdevs_list": [ 00:34:44.020 { 00:34:44.020 "name": null, 00:34:44.020 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:44.020 "is_configured": false, 00:34:44.020 "data_offset": 0, 00:34:44.020 "data_size": 7936 00:34:44.020 }, 00:34:44.020 { 00:34:44.020 "name": "BaseBdev2", 00:34:44.020 "uuid": "462c1e65-ee53-50fd-b426-a46b5461db00", 00:34:44.020 "is_configured": true, 00:34:44.020 "data_offset": 256, 00:34:44.020 "data_size": 7936 00:34:44.020 } 00:34:44.020 ] 00:34:44.020 }' 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:44.020 13:54:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:44.279 13:54:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:34:44.279 13:54:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:44.279 13:54:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:44.279 [2024-11-20 13:54:51.297114] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:34:44.279 [2024-11-20 13:54:51.307218] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00018d260 00:34:44.279 13:54:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:44.279 13:54:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@647 -- # sleep 1 00:34:44.279 [2024-11-20 13:54:51.308890] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:45.655 "name": "raid_bdev1", 00:34:45.655 "uuid": "86cc1961-187e-4be6-b876-3fdac7636128", 00:34:45.655 "strip_size_kb": 0, 00:34:45.655 "state": "online", 00:34:45.655 "raid_level": "raid1", 00:34:45.655 "superblock": true, 00:34:45.655 "num_base_bdevs": 2, 00:34:45.655 "num_base_bdevs_discovered": 2, 00:34:45.655 "num_base_bdevs_operational": 2, 00:34:45.655 "process": { 00:34:45.655 "type": "rebuild", 00:34:45.655 "target": "spare", 00:34:45.655 "progress": { 00:34:45.655 "blocks": 2560, 00:34:45.655 "percent": 32 00:34:45.655 } 00:34:45.655 }, 00:34:45.655 "base_bdevs_list": [ 00:34:45.655 { 00:34:45.655 "name": "spare", 00:34:45.655 "uuid": "074342c0-5eb6-58be-bac4-f3d582af5186", 00:34:45.655 "is_configured": true, 00:34:45.655 "data_offset": 256, 00:34:45.655 "data_size": 7936 00:34:45.655 }, 00:34:45.655 { 00:34:45.655 "name": "BaseBdev2", 00:34:45.655 "uuid": "462c1e65-ee53-50fd-b426-a46b5461db00", 00:34:45.655 "is_configured": true, 00:34:45.655 "data_offset": 256, 00:34:45.655 "data_size": 7936 00:34:45.655 } 00:34:45.655 ] 00:34:45.655 }' 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:45.655 [2024-11-20 13:54:52.419036] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:34:45.655 [2024-11-20 13:54:52.516158] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:34:45.655 [2024-11-20 13:54:52.516381] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:34:45.655 [2024-11-20 13:54:52.516398] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:34:45.655 [2024-11-20 13:54:52.516411] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:45.655 "name": "raid_bdev1", 00:34:45.655 "uuid": "86cc1961-187e-4be6-b876-3fdac7636128", 00:34:45.655 "strip_size_kb": 0, 00:34:45.655 "state": "online", 00:34:45.655 "raid_level": "raid1", 00:34:45.655 "superblock": true, 00:34:45.655 "num_base_bdevs": 2, 00:34:45.655 "num_base_bdevs_discovered": 1, 00:34:45.655 "num_base_bdevs_operational": 1, 00:34:45.655 "base_bdevs_list": [ 00:34:45.655 { 00:34:45.655 "name": null, 00:34:45.655 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:45.655 "is_configured": false, 00:34:45.655 "data_offset": 0, 00:34:45.655 "data_size": 7936 00:34:45.655 }, 00:34:45.655 { 00:34:45.655 "name": "BaseBdev2", 00:34:45.655 "uuid": "462c1e65-ee53-50fd-b426-a46b5461db00", 00:34:45.655 "is_configured": true, 00:34:45.655 "data_offset": 256, 00:34:45.655 "data_size": 7936 00:34:45.655 } 00:34:45.655 ] 00:34:45.655 }' 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:45.655 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:45.914 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:34:45.914 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:45.914 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:34:45.914 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:34:45.914 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:45.914 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:45.914 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:45.914 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:45.914 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:45.914 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:45.914 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:45.914 "name": "raid_bdev1", 00:34:45.914 "uuid": "86cc1961-187e-4be6-b876-3fdac7636128", 00:34:45.914 "strip_size_kb": 0, 00:34:45.914 "state": "online", 00:34:45.914 "raid_level": "raid1", 00:34:45.914 "superblock": true, 00:34:45.914 "num_base_bdevs": 2, 00:34:45.914 "num_base_bdevs_discovered": 1, 00:34:45.914 "num_base_bdevs_operational": 1, 00:34:45.914 "base_bdevs_list": [ 00:34:45.914 { 00:34:45.914 "name": null, 00:34:45.914 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:45.914 "is_configured": false, 00:34:45.914 "data_offset": 0, 00:34:45.914 "data_size": 7936 00:34:45.914 }, 00:34:45.914 { 00:34:45.914 "name": "BaseBdev2", 00:34:45.914 "uuid": "462c1e65-ee53-50fd-b426-a46b5461db00", 00:34:45.914 "is_configured": true, 00:34:45.914 "data_offset": 256, 00:34:45.914 "data_size": 7936 00:34:45.914 } 00:34:45.914 ] 00:34:45.914 }' 00:34:45.914 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:45.914 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:34:45.914 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:45.914 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:34:45.914 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:34:45.914 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:45.914 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:46.173 [2024-11-20 13:54:52.972163] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:34:46.173 [2024-11-20 13:54:52.981765] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00018d330 00:34:46.173 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:46.173 13:54:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@663 -- # sleep 1 00:34:46.173 [2024-11-20 13:54:52.983445] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:34:47.144 13:54:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:47.144 13:54:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:47.144 13:54:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:47.144 13:54:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:47.144 13:54:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:47.144 13:54:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:47.144 13:54:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:47.144 13:54:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:47.144 13:54:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:47.144 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:47.144 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:47.144 "name": "raid_bdev1", 00:34:47.144 "uuid": "86cc1961-187e-4be6-b876-3fdac7636128", 00:34:47.144 "strip_size_kb": 0, 00:34:47.144 "state": "online", 00:34:47.144 "raid_level": "raid1", 00:34:47.144 "superblock": true, 00:34:47.144 "num_base_bdevs": 2, 00:34:47.144 "num_base_bdevs_discovered": 2, 00:34:47.144 "num_base_bdevs_operational": 2, 00:34:47.144 "process": { 00:34:47.144 "type": "rebuild", 00:34:47.144 "target": "spare", 00:34:47.144 "progress": { 00:34:47.144 "blocks": 2560, 00:34:47.144 "percent": 32 00:34:47.144 } 00:34:47.144 }, 00:34:47.144 "base_bdevs_list": [ 00:34:47.144 { 00:34:47.144 "name": "spare", 00:34:47.144 "uuid": "074342c0-5eb6-58be-bac4-f3d582af5186", 00:34:47.144 "is_configured": true, 00:34:47.144 "data_offset": 256, 00:34:47.144 "data_size": 7936 00:34:47.144 }, 00:34:47.144 { 00:34:47.144 "name": "BaseBdev2", 00:34:47.144 "uuid": "462c1e65-ee53-50fd-b426-a46b5461db00", 00:34:47.144 "is_configured": true, 00:34:47.144 "data_offset": 256, 00:34:47.144 "data_size": 7936 00:34:47.144 } 00:34:47.144 ] 00:34:47.144 }' 00:34:47.144 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:47.144 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:34:47.144 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:47.144 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:34:47.144 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:34:47.144 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:34:47.144 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:34:47.144 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:34:47.144 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:34:47.144 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:34:47.144 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@706 -- # local timeout=541 00:34:47.144 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:34:47.144 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:47.144 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:47.144 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:47.144 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:47.144 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:47.144 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:47.144 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:47.144 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:47.144 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:47.144 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:47.144 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:47.144 "name": "raid_bdev1", 00:34:47.144 "uuid": "86cc1961-187e-4be6-b876-3fdac7636128", 00:34:47.144 "strip_size_kb": 0, 00:34:47.144 "state": "online", 00:34:47.144 "raid_level": "raid1", 00:34:47.144 "superblock": true, 00:34:47.144 "num_base_bdevs": 2, 00:34:47.144 "num_base_bdevs_discovered": 2, 00:34:47.144 "num_base_bdevs_operational": 2, 00:34:47.144 "process": { 00:34:47.144 "type": "rebuild", 00:34:47.144 "target": "spare", 00:34:47.145 "progress": { 00:34:47.145 "blocks": 2560, 00:34:47.145 "percent": 32 00:34:47.145 } 00:34:47.145 }, 00:34:47.145 "base_bdevs_list": [ 00:34:47.145 { 00:34:47.145 "name": "spare", 00:34:47.145 "uuid": "074342c0-5eb6-58be-bac4-f3d582af5186", 00:34:47.145 "is_configured": true, 00:34:47.145 "data_offset": 256, 00:34:47.145 "data_size": 7936 00:34:47.145 }, 00:34:47.145 { 00:34:47.145 "name": "BaseBdev2", 00:34:47.145 "uuid": "462c1e65-ee53-50fd-b426-a46b5461db00", 00:34:47.145 "is_configured": true, 00:34:47.145 "data_offset": 256, 00:34:47.145 "data_size": 7936 00:34:47.145 } 00:34:47.145 ] 00:34:47.145 }' 00:34:47.145 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:47.145 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:34:47.145 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:47.145 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:34:47.145 13:54:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@711 -- # sleep 1 00:34:48.518 13:54:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:34:48.518 13:54:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:48.518 13:54:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:48.518 13:54:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:48.518 13:54:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:48.518 13:54:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:48.518 13:54:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:48.518 13:54:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:48.518 13:54:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:48.518 13:54:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:48.518 13:54:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:48.518 13:54:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:48.518 "name": "raid_bdev1", 00:34:48.518 "uuid": "86cc1961-187e-4be6-b876-3fdac7636128", 00:34:48.518 "strip_size_kb": 0, 00:34:48.518 "state": "online", 00:34:48.518 "raid_level": "raid1", 00:34:48.518 "superblock": true, 00:34:48.518 "num_base_bdevs": 2, 00:34:48.518 "num_base_bdevs_discovered": 2, 00:34:48.518 "num_base_bdevs_operational": 2, 00:34:48.518 "process": { 00:34:48.518 "type": "rebuild", 00:34:48.518 "target": "spare", 00:34:48.518 "progress": { 00:34:48.518 "blocks": 5376, 00:34:48.518 "percent": 67 00:34:48.518 } 00:34:48.518 }, 00:34:48.518 "base_bdevs_list": [ 00:34:48.518 { 00:34:48.518 "name": "spare", 00:34:48.518 "uuid": "074342c0-5eb6-58be-bac4-f3d582af5186", 00:34:48.518 "is_configured": true, 00:34:48.518 "data_offset": 256, 00:34:48.518 "data_size": 7936 00:34:48.518 }, 00:34:48.518 { 00:34:48.518 "name": "BaseBdev2", 00:34:48.518 "uuid": "462c1e65-ee53-50fd-b426-a46b5461db00", 00:34:48.518 "is_configured": true, 00:34:48.518 "data_offset": 256, 00:34:48.518 "data_size": 7936 00:34:48.518 } 00:34:48.518 ] 00:34:48.518 }' 00:34:48.518 13:54:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:48.518 13:54:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:34:48.518 13:54:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:48.518 13:54:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:34:48.518 13:54:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@711 -- # sleep 1 00:34:49.083 [2024-11-20 13:54:56.101404] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:34:49.083 [2024-11-20 13:54:56.101487] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:34:49.083 [2024-11-20 13:54:56.101596] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:49.341 "name": "raid_bdev1", 00:34:49.341 "uuid": "86cc1961-187e-4be6-b876-3fdac7636128", 00:34:49.341 "strip_size_kb": 0, 00:34:49.341 "state": "online", 00:34:49.341 "raid_level": "raid1", 00:34:49.341 "superblock": true, 00:34:49.341 "num_base_bdevs": 2, 00:34:49.341 "num_base_bdevs_discovered": 2, 00:34:49.341 "num_base_bdevs_operational": 2, 00:34:49.341 "base_bdevs_list": [ 00:34:49.341 { 00:34:49.341 "name": "spare", 00:34:49.341 "uuid": "074342c0-5eb6-58be-bac4-f3d582af5186", 00:34:49.341 "is_configured": true, 00:34:49.341 "data_offset": 256, 00:34:49.341 "data_size": 7936 00:34:49.341 }, 00:34:49.341 { 00:34:49.341 "name": "BaseBdev2", 00:34:49.341 "uuid": "462c1e65-ee53-50fd-b426-a46b5461db00", 00:34:49.341 "is_configured": true, 00:34:49.341 "data_offset": 256, 00:34:49.341 "data_size": 7936 00:34:49.341 } 00:34:49.341 ] 00:34:49.341 }' 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@709 -- # break 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:49.341 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:49.599 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:49.599 "name": "raid_bdev1", 00:34:49.599 "uuid": "86cc1961-187e-4be6-b876-3fdac7636128", 00:34:49.599 "strip_size_kb": 0, 00:34:49.599 "state": "online", 00:34:49.599 "raid_level": "raid1", 00:34:49.599 "superblock": true, 00:34:49.599 "num_base_bdevs": 2, 00:34:49.599 "num_base_bdevs_discovered": 2, 00:34:49.599 "num_base_bdevs_operational": 2, 00:34:49.599 "base_bdevs_list": [ 00:34:49.599 { 00:34:49.599 "name": "spare", 00:34:49.599 "uuid": "074342c0-5eb6-58be-bac4-f3d582af5186", 00:34:49.599 "is_configured": true, 00:34:49.599 "data_offset": 256, 00:34:49.599 "data_size": 7936 00:34:49.599 }, 00:34:49.599 { 00:34:49.599 "name": "BaseBdev2", 00:34:49.599 "uuid": "462c1e65-ee53-50fd-b426-a46b5461db00", 00:34:49.599 "is_configured": true, 00:34:49.599 "data_offset": 256, 00:34:49.599 "data_size": 7936 00:34:49.599 } 00:34:49.599 ] 00:34:49.599 }' 00:34:49.599 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:49.599 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:34:49.599 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:49.599 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:34:49.599 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:34:49.599 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:49.599 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:49.599 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:34:49.599 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:34:49.599 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:34:49.599 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:49.599 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:49.599 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:49.599 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:49.599 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:49.599 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:49.599 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:49.599 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:49.599 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:49.599 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:49.599 "name": "raid_bdev1", 00:34:49.599 "uuid": "86cc1961-187e-4be6-b876-3fdac7636128", 00:34:49.599 "strip_size_kb": 0, 00:34:49.599 "state": "online", 00:34:49.599 "raid_level": "raid1", 00:34:49.599 "superblock": true, 00:34:49.599 "num_base_bdevs": 2, 00:34:49.599 "num_base_bdevs_discovered": 2, 00:34:49.599 "num_base_bdevs_operational": 2, 00:34:49.599 "base_bdevs_list": [ 00:34:49.599 { 00:34:49.599 "name": "spare", 00:34:49.599 "uuid": "074342c0-5eb6-58be-bac4-f3d582af5186", 00:34:49.599 "is_configured": true, 00:34:49.599 "data_offset": 256, 00:34:49.599 "data_size": 7936 00:34:49.599 }, 00:34:49.600 { 00:34:49.600 "name": "BaseBdev2", 00:34:49.600 "uuid": "462c1e65-ee53-50fd-b426-a46b5461db00", 00:34:49.600 "is_configured": true, 00:34:49.600 "data_offset": 256, 00:34:49.600 "data_size": 7936 00:34:49.600 } 00:34:49.600 ] 00:34:49.600 }' 00:34:49.600 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:49.600 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:49.857 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:34:49.857 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:49.857 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:49.857 [2024-11-20 13:54:56.804956] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:34:49.857 [2024-11-20 13:54:56.805026] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:34:49.857 [2024-11-20 13:54:56.805124] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:34:49.857 [2024-11-20 13:54:56.805192] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:34:49.857 [2024-11-20 13:54:56.805204] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:34:49.857 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:49.857 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # jq length 00:34:49.857 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:49.857 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:49.857 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:49.857 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:49.857 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:34:49.857 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:34:49.857 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:34:49.857 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:34:49.857 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:34:49.857 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:34:49.857 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # local bdev_list 00:34:49.857 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:34:49.857 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # local nbd_list 00:34:49.857 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@12 -- # local i 00:34:49.857 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:34:49.857 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:34:49.857 13:54:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:34:50.115 /dev/nbd0 00:34:50.115 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:34:50.115 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:34:50.115 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:34:50.115 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # local i 00:34:50.115 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:34:50.115 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:34:50.115 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:34:50.115 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@877 -- # break 00:34:50.115 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:34:50.115 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:34:50.115 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:34:50.115 1+0 records in 00:34:50.115 1+0 records out 00:34:50.115 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000232757 s, 17.6 MB/s 00:34:50.115 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:34:50.115 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # size=4096 00:34:50.115 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:34:50.115 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:34:50.115 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@893 -- # return 0 00:34:50.115 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:34:50.115 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:34:50.115 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:34:50.448 /dev/nbd1 00:34:50.448 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:34:50.448 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:34:50.448 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:34:50.448 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # local i 00:34:50.448 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:34:50.448 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:34:50.448 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:34:50.448 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@877 -- # break 00:34:50.448 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:34:50.448 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:34:50.448 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:34:50.448 1+0 records in 00:34:50.448 1+0 records out 00:34:50.448 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000307235 s, 13.3 MB/s 00:34:50.448 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:34:50.448 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # size=4096 00:34:50.448 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:34:50.448 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:34:50.448 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@893 -- # return 0 00:34:50.448 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:34:50.448 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:34:50.448 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:34:50.448 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:34:50.448 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:34:50.448 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:34:50.448 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # local nbd_list 00:34:50.448 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@51 -- # local i 00:34:50.448 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:34:50.448 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:34:50.706 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:34:50.706 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:34:50.706 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:34:50.706 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:34:50.706 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:34:50.706 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:34:50.706 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:34:50.706 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:34:50.706 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:34:50.706 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:50.965 [2024-11-20 13:54:57.859178] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:34:50.965 [2024-11-20 13:54:57.859242] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:34:50.965 [2024-11-20 13:54:57.859267] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:34:50.965 [2024-11-20 13:54:57.859276] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:34:50.965 [2024-11-20 13:54:57.861329] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:34:50.965 [2024-11-20 13:54:57.861363] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:34:50.965 [2024-11-20 13:54:57.861457] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:34:50.965 [2024-11-20 13:54:57.861507] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:34:50.965 [2024-11-20 13:54:57.861633] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:34:50.965 spare 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:50.965 [2024-11-20 13:54:57.961733] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:34:50.965 [2024-11-20 13:54:57.961792] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:34:50.965 [2024-11-20 13:54:57.962135] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1b50 00:34:50.965 [2024-11-20 13:54:57.962334] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:34:50.965 [2024-11-20 13:54:57.962349] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:34:50.965 [2024-11-20 13:54:57.962523] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:50.965 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:50.966 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:50.966 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:50.966 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:50.966 "name": "raid_bdev1", 00:34:50.966 "uuid": "86cc1961-187e-4be6-b876-3fdac7636128", 00:34:50.966 "strip_size_kb": 0, 00:34:50.966 "state": "online", 00:34:50.966 "raid_level": "raid1", 00:34:50.966 "superblock": true, 00:34:50.966 "num_base_bdevs": 2, 00:34:50.966 "num_base_bdevs_discovered": 2, 00:34:50.966 "num_base_bdevs_operational": 2, 00:34:50.966 "base_bdevs_list": [ 00:34:50.966 { 00:34:50.966 "name": "spare", 00:34:50.966 "uuid": "074342c0-5eb6-58be-bac4-f3d582af5186", 00:34:50.966 "is_configured": true, 00:34:50.966 "data_offset": 256, 00:34:50.966 "data_size": 7936 00:34:50.966 }, 00:34:50.966 { 00:34:50.966 "name": "BaseBdev2", 00:34:50.966 "uuid": "462c1e65-ee53-50fd-b426-a46b5461db00", 00:34:50.966 "is_configured": true, 00:34:50.966 "data_offset": 256, 00:34:50.966 "data_size": 7936 00:34:50.966 } 00:34:50.966 ] 00:34:50.966 }' 00:34:50.966 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:50.966 13:54:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:51.532 "name": "raid_bdev1", 00:34:51.532 "uuid": "86cc1961-187e-4be6-b876-3fdac7636128", 00:34:51.532 "strip_size_kb": 0, 00:34:51.532 "state": "online", 00:34:51.532 "raid_level": "raid1", 00:34:51.532 "superblock": true, 00:34:51.532 "num_base_bdevs": 2, 00:34:51.532 "num_base_bdevs_discovered": 2, 00:34:51.532 "num_base_bdevs_operational": 2, 00:34:51.532 "base_bdevs_list": [ 00:34:51.532 { 00:34:51.532 "name": "spare", 00:34:51.532 "uuid": "074342c0-5eb6-58be-bac4-f3d582af5186", 00:34:51.532 "is_configured": true, 00:34:51.532 "data_offset": 256, 00:34:51.532 "data_size": 7936 00:34:51.532 }, 00:34:51.532 { 00:34:51.532 "name": "BaseBdev2", 00:34:51.532 "uuid": "462c1e65-ee53-50fd-b426-a46b5461db00", 00:34:51.532 "is_configured": true, 00:34:51.532 "data_offset": 256, 00:34:51.532 "data_size": 7936 00:34:51.532 } 00:34:51.532 ] 00:34:51.532 }' 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:51.532 [2024-11-20 13:54:58.435361] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:51.532 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:51.533 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:34:51.533 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:34:51.533 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:34:51.533 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:51.533 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:51.533 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:51.533 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:51.533 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:51.533 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:51.533 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:51.533 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:51.533 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:51.533 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:51.533 "name": "raid_bdev1", 00:34:51.533 "uuid": "86cc1961-187e-4be6-b876-3fdac7636128", 00:34:51.533 "strip_size_kb": 0, 00:34:51.533 "state": "online", 00:34:51.533 "raid_level": "raid1", 00:34:51.533 "superblock": true, 00:34:51.533 "num_base_bdevs": 2, 00:34:51.533 "num_base_bdevs_discovered": 1, 00:34:51.533 "num_base_bdevs_operational": 1, 00:34:51.533 "base_bdevs_list": [ 00:34:51.533 { 00:34:51.533 "name": null, 00:34:51.533 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:51.533 "is_configured": false, 00:34:51.533 "data_offset": 0, 00:34:51.533 "data_size": 7936 00:34:51.533 }, 00:34:51.533 { 00:34:51.533 "name": "BaseBdev2", 00:34:51.533 "uuid": "462c1e65-ee53-50fd-b426-a46b5461db00", 00:34:51.533 "is_configured": true, 00:34:51.533 "data_offset": 256, 00:34:51.533 "data_size": 7936 00:34:51.533 } 00:34:51.533 ] 00:34:51.533 }' 00:34:51.533 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:51.533 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:51.791 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:34:51.791 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:51.791 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:51.791 [2024-11-20 13:54:58.759419] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:34:51.791 [2024-11-20 13:54:58.759620] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:34:51.791 [2024-11-20 13:54:58.759635] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:34:51.791 [2024-11-20 13:54:58.759669] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:34:51.791 [2024-11-20 13:54:58.769156] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1c20 00:34:51.791 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:51.791 13:54:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@757 -- # sleep 1 00:34:51.791 [2024-11-20 13:54:58.770842] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:34:52.723 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:52.723 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:52.723 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:52.723 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:52.723 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:52.723 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:52.723 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:52.723 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:52.723 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:52.979 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:52.979 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:52.979 "name": "raid_bdev1", 00:34:52.979 "uuid": "86cc1961-187e-4be6-b876-3fdac7636128", 00:34:52.979 "strip_size_kb": 0, 00:34:52.979 "state": "online", 00:34:52.979 "raid_level": "raid1", 00:34:52.979 "superblock": true, 00:34:52.979 "num_base_bdevs": 2, 00:34:52.979 "num_base_bdevs_discovered": 2, 00:34:52.979 "num_base_bdevs_operational": 2, 00:34:52.979 "process": { 00:34:52.979 "type": "rebuild", 00:34:52.979 "target": "spare", 00:34:52.979 "progress": { 00:34:52.979 "blocks": 2560, 00:34:52.979 "percent": 32 00:34:52.979 } 00:34:52.979 }, 00:34:52.979 "base_bdevs_list": [ 00:34:52.979 { 00:34:52.979 "name": "spare", 00:34:52.979 "uuid": "074342c0-5eb6-58be-bac4-f3d582af5186", 00:34:52.979 "is_configured": true, 00:34:52.979 "data_offset": 256, 00:34:52.979 "data_size": 7936 00:34:52.979 }, 00:34:52.979 { 00:34:52.979 "name": "BaseBdev2", 00:34:52.979 "uuid": "462c1e65-ee53-50fd-b426-a46b5461db00", 00:34:52.979 "is_configured": true, 00:34:52.979 "data_offset": 256, 00:34:52.979 "data_size": 7936 00:34:52.979 } 00:34:52.979 ] 00:34:52.979 }' 00:34:52.979 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:52.979 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:34:52.979 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:52.979 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:34:52.979 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:34:52.979 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:52.979 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:52.979 [2024-11-20 13:54:59.884950] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:34:52.979 [2024-11-20 13:54:59.977990] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:34:52.979 [2024-11-20 13:54:59.978072] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:34:52.979 [2024-11-20 13:54:59.978086] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:34:52.979 [2024-11-20 13:54:59.978095] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:34:52.979 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:52.979 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:34:52.979 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:52.979 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:52.979 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:34:52.979 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:34:52.979 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:34:52.979 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:52.979 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:52.979 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:52.979 13:54:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:52.979 13:55:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:52.979 13:55:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:52.979 13:55:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:52.979 13:55:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:52.979 13:55:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:52.979 13:55:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:52.979 "name": "raid_bdev1", 00:34:52.979 "uuid": "86cc1961-187e-4be6-b876-3fdac7636128", 00:34:52.979 "strip_size_kb": 0, 00:34:52.979 "state": "online", 00:34:52.979 "raid_level": "raid1", 00:34:52.979 "superblock": true, 00:34:52.979 "num_base_bdevs": 2, 00:34:52.979 "num_base_bdevs_discovered": 1, 00:34:52.979 "num_base_bdevs_operational": 1, 00:34:52.979 "base_bdevs_list": [ 00:34:52.979 { 00:34:52.979 "name": null, 00:34:52.979 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:52.979 "is_configured": false, 00:34:52.979 "data_offset": 0, 00:34:52.979 "data_size": 7936 00:34:52.979 }, 00:34:52.979 { 00:34:52.979 "name": "BaseBdev2", 00:34:52.979 "uuid": "462c1e65-ee53-50fd-b426-a46b5461db00", 00:34:52.979 "is_configured": true, 00:34:52.979 "data_offset": 256, 00:34:52.979 "data_size": 7936 00:34:52.979 } 00:34:52.979 ] 00:34:52.979 }' 00:34:52.979 13:55:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:52.979 13:55:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:53.545 13:55:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:34:53.545 13:55:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:53.545 13:55:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:53.545 [2024-11-20 13:55:00.301503] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:34:53.545 [2024-11-20 13:55:00.301571] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:34:53.545 [2024-11-20 13:55:00.301593] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:34:53.545 [2024-11-20 13:55:00.301603] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:34:53.545 [2024-11-20 13:55:00.302062] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:34:53.545 [2024-11-20 13:55:00.302085] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:34:53.545 [2024-11-20 13:55:00.302174] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:34:53.545 [2024-11-20 13:55:00.302188] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:34:53.545 [2024-11-20 13:55:00.302200] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:34:53.545 [2024-11-20 13:55:00.302225] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:34:53.545 [2024-11-20 13:55:00.311575] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1cf0 00:34:53.545 spare 00:34:53.545 13:55:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:53.545 13:55:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@764 -- # sleep 1 00:34:53.545 [2024-11-20 13:55:00.313258] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:54.478 "name": "raid_bdev1", 00:34:54.478 "uuid": "86cc1961-187e-4be6-b876-3fdac7636128", 00:34:54.478 "strip_size_kb": 0, 00:34:54.478 "state": "online", 00:34:54.478 "raid_level": "raid1", 00:34:54.478 "superblock": true, 00:34:54.478 "num_base_bdevs": 2, 00:34:54.478 "num_base_bdevs_discovered": 2, 00:34:54.478 "num_base_bdevs_operational": 2, 00:34:54.478 "process": { 00:34:54.478 "type": "rebuild", 00:34:54.478 "target": "spare", 00:34:54.478 "progress": { 00:34:54.478 "blocks": 2560, 00:34:54.478 "percent": 32 00:34:54.478 } 00:34:54.478 }, 00:34:54.478 "base_bdevs_list": [ 00:34:54.478 { 00:34:54.478 "name": "spare", 00:34:54.478 "uuid": "074342c0-5eb6-58be-bac4-f3d582af5186", 00:34:54.478 "is_configured": true, 00:34:54.478 "data_offset": 256, 00:34:54.478 "data_size": 7936 00:34:54.478 }, 00:34:54.478 { 00:34:54.478 "name": "BaseBdev2", 00:34:54.478 "uuid": "462c1e65-ee53-50fd-b426-a46b5461db00", 00:34:54.478 "is_configured": true, 00:34:54.478 "data_offset": 256, 00:34:54.478 "data_size": 7936 00:34:54.478 } 00:34:54.478 ] 00:34:54.478 }' 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:54.478 [2024-11-20 13:55:01.415761] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:34:54.478 [2024-11-20 13:55:01.419674] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:34:54.478 [2024-11-20 13:55:01.419725] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:34:54.478 [2024-11-20 13:55:01.419740] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:34:54.478 [2024-11-20 13:55:01.419747] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:54.478 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:54.478 "name": "raid_bdev1", 00:34:54.478 "uuid": "86cc1961-187e-4be6-b876-3fdac7636128", 00:34:54.478 "strip_size_kb": 0, 00:34:54.478 "state": "online", 00:34:54.478 "raid_level": "raid1", 00:34:54.478 "superblock": true, 00:34:54.478 "num_base_bdevs": 2, 00:34:54.478 "num_base_bdevs_discovered": 1, 00:34:54.478 "num_base_bdevs_operational": 1, 00:34:54.478 "base_bdevs_list": [ 00:34:54.478 { 00:34:54.478 "name": null, 00:34:54.479 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:54.479 "is_configured": false, 00:34:54.479 "data_offset": 0, 00:34:54.479 "data_size": 7936 00:34:54.479 }, 00:34:54.479 { 00:34:54.479 "name": "BaseBdev2", 00:34:54.479 "uuid": "462c1e65-ee53-50fd-b426-a46b5461db00", 00:34:54.479 "is_configured": true, 00:34:54.479 "data_offset": 256, 00:34:54.479 "data_size": 7936 00:34:54.479 } 00:34:54.479 ] 00:34:54.479 }' 00:34:54.479 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:54.479 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:54.737 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:34:54.737 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:54.737 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:34:54.737 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:34:54.737 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:54.737 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:54.737 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:54.737 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:54.737 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:54.737 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:54.737 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:54.737 "name": "raid_bdev1", 00:34:54.737 "uuid": "86cc1961-187e-4be6-b876-3fdac7636128", 00:34:54.737 "strip_size_kb": 0, 00:34:54.737 "state": "online", 00:34:54.737 "raid_level": "raid1", 00:34:54.737 "superblock": true, 00:34:54.737 "num_base_bdevs": 2, 00:34:54.737 "num_base_bdevs_discovered": 1, 00:34:54.737 "num_base_bdevs_operational": 1, 00:34:54.737 "base_bdevs_list": [ 00:34:54.737 { 00:34:54.737 "name": null, 00:34:54.737 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:54.737 "is_configured": false, 00:34:54.737 "data_offset": 0, 00:34:54.737 "data_size": 7936 00:34:54.737 }, 00:34:54.737 { 00:34:54.737 "name": "BaseBdev2", 00:34:54.737 "uuid": "462c1e65-ee53-50fd-b426-a46b5461db00", 00:34:54.737 "is_configured": true, 00:34:54.737 "data_offset": 256, 00:34:54.737 "data_size": 7936 00:34:54.737 } 00:34:54.737 ] 00:34:54.737 }' 00:34:54.737 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:54.994 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:34:54.994 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:54.994 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:34:54.994 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:34:54.994 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:54.994 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:54.994 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:54.994 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:34:54.994 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:54.994 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:54.994 [2024-11-20 13:55:01.863221] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:34:54.994 [2024-11-20 13:55:01.863280] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:34:54.994 [2024-11-20 13:55:01.863305] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:34:54.994 [2024-11-20 13:55:01.863313] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:34:54.994 [2024-11-20 13:55:01.863740] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:34:54.994 [2024-11-20 13:55:01.863758] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:34:54.994 [2024-11-20 13:55:01.863831] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:34:54.994 [2024-11-20 13:55:01.863844] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:34:54.994 [2024-11-20 13:55:01.863852] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:34:54.994 [2024-11-20 13:55:01.863861] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:34:54.994 BaseBdev1 00:34:54.994 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:54.994 13:55:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@775 -- # sleep 1 00:34:55.927 13:55:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:34:55.927 13:55:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:55.927 13:55:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:55.927 13:55:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:34:55.927 13:55:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:34:55.927 13:55:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:34:55.927 13:55:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:55.927 13:55:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:55.927 13:55:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:55.927 13:55:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:55.927 13:55:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:55.927 13:55:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:55.927 13:55:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:55.927 13:55:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:55.927 13:55:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:55.927 13:55:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:55.927 "name": "raid_bdev1", 00:34:55.927 "uuid": "86cc1961-187e-4be6-b876-3fdac7636128", 00:34:55.927 "strip_size_kb": 0, 00:34:55.927 "state": "online", 00:34:55.927 "raid_level": "raid1", 00:34:55.927 "superblock": true, 00:34:55.927 "num_base_bdevs": 2, 00:34:55.927 "num_base_bdevs_discovered": 1, 00:34:55.927 "num_base_bdevs_operational": 1, 00:34:55.927 "base_bdevs_list": [ 00:34:55.927 { 00:34:55.927 "name": null, 00:34:55.927 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:55.927 "is_configured": false, 00:34:55.927 "data_offset": 0, 00:34:55.927 "data_size": 7936 00:34:55.927 }, 00:34:55.927 { 00:34:55.927 "name": "BaseBdev2", 00:34:55.927 "uuid": "462c1e65-ee53-50fd-b426-a46b5461db00", 00:34:55.927 "is_configured": true, 00:34:55.927 "data_offset": 256, 00:34:55.927 "data_size": 7936 00:34:55.927 } 00:34:55.927 ] 00:34:55.927 }' 00:34:55.927 13:55:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:55.927 13:55:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:56.185 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:34:56.185 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:56.185 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:34:56.185 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:34:56.185 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:56.185 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:56.185 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:56.185 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:56.185 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:56.185 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:56.185 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:56.185 "name": "raid_bdev1", 00:34:56.185 "uuid": "86cc1961-187e-4be6-b876-3fdac7636128", 00:34:56.185 "strip_size_kb": 0, 00:34:56.185 "state": "online", 00:34:56.185 "raid_level": "raid1", 00:34:56.185 "superblock": true, 00:34:56.185 "num_base_bdevs": 2, 00:34:56.185 "num_base_bdevs_discovered": 1, 00:34:56.185 "num_base_bdevs_operational": 1, 00:34:56.185 "base_bdevs_list": [ 00:34:56.185 { 00:34:56.185 "name": null, 00:34:56.185 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:56.185 "is_configured": false, 00:34:56.185 "data_offset": 0, 00:34:56.185 "data_size": 7936 00:34:56.185 }, 00:34:56.185 { 00:34:56.185 "name": "BaseBdev2", 00:34:56.185 "uuid": "462c1e65-ee53-50fd-b426-a46b5461db00", 00:34:56.185 "is_configured": true, 00:34:56.185 "data_offset": 256, 00:34:56.185 "data_size": 7936 00:34:56.185 } 00:34:56.185 ] 00:34:56.185 }' 00:34:56.185 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:56.443 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:34:56.443 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:56.443 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:34:56.443 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:34:56.443 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@652 -- # local es=0 00:34:56.443 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:34:56.443 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:34:56.443 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:34:56.443 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:34:56.443 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:34:56.443 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:34:56.443 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:56.443 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:56.443 [2024-11-20 13:55:03.291566] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:34:56.443 [2024-11-20 13:55:03.291733] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:34:56.443 [2024-11-20 13:55:03.291753] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:34:56.443 request: 00:34:56.443 { 00:34:56.443 "base_bdev": "BaseBdev1", 00:34:56.443 "raid_bdev": "raid_bdev1", 00:34:56.443 "method": "bdev_raid_add_base_bdev", 00:34:56.443 "req_id": 1 00:34:56.443 } 00:34:56.443 Got JSON-RPC error response 00:34:56.443 response: 00:34:56.443 { 00:34:56.443 "code": -22, 00:34:56.443 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:34:56.443 } 00:34:56.443 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:34:56.443 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@655 -- # es=1 00:34:56.443 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:34:56.443 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:34:56.443 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:34:56.443 13:55:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@779 -- # sleep 1 00:34:57.378 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:34:57.378 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:34:57.378 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:34:57.378 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:34:57.378 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:34:57.378 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:34:57.378 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:57.378 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:57.378 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:57.378 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:57.378 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:57.378 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:57.378 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:57.378 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:57.378 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:57.378 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:57.378 "name": "raid_bdev1", 00:34:57.378 "uuid": "86cc1961-187e-4be6-b876-3fdac7636128", 00:34:57.378 "strip_size_kb": 0, 00:34:57.378 "state": "online", 00:34:57.378 "raid_level": "raid1", 00:34:57.378 "superblock": true, 00:34:57.378 "num_base_bdevs": 2, 00:34:57.378 "num_base_bdevs_discovered": 1, 00:34:57.378 "num_base_bdevs_operational": 1, 00:34:57.378 "base_bdevs_list": [ 00:34:57.378 { 00:34:57.378 "name": null, 00:34:57.378 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:57.378 "is_configured": false, 00:34:57.378 "data_offset": 0, 00:34:57.378 "data_size": 7936 00:34:57.378 }, 00:34:57.378 { 00:34:57.378 "name": "BaseBdev2", 00:34:57.378 "uuid": "462c1e65-ee53-50fd-b426-a46b5461db00", 00:34:57.378 "is_configured": true, 00:34:57.378 "data_offset": 256, 00:34:57.378 "data_size": 7936 00:34:57.378 } 00:34:57.378 ] 00:34:57.378 }' 00:34:57.378 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:57.378 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:57.636 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:34:57.636 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:34:57.636 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:34:57.636 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:34:57.636 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:34:57.636 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:57.636 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:57.636 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:57.636 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:34:57.636 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:57.636 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:34:57.636 "name": "raid_bdev1", 00:34:57.636 "uuid": "86cc1961-187e-4be6-b876-3fdac7636128", 00:34:57.636 "strip_size_kb": 0, 00:34:57.636 "state": "online", 00:34:57.636 "raid_level": "raid1", 00:34:57.636 "superblock": true, 00:34:57.636 "num_base_bdevs": 2, 00:34:57.636 "num_base_bdevs_discovered": 1, 00:34:57.636 "num_base_bdevs_operational": 1, 00:34:57.636 "base_bdevs_list": [ 00:34:57.636 { 00:34:57.636 "name": null, 00:34:57.636 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:57.636 "is_configured": false, 00:34:57.636 "data_offset": 0, 00:34:57.636 "data_size": 7936 00:34:57.636 }, 00:34:57.636 { 00:34:57.636 "name": "BaseBdev2", 00:34:57.636 "uuid": "462c1e65-ee53-50fd-b426-a46b5461db00", 00:34:57.636 "is_configured": true, 00:34:57.636 "data_offset": 256, 00:34:57.636 "data_size": 7936 00:34:57.636 } 00:34:57.636 ] 00:34:57.636 }' 00:34:57.636 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:34:57.636 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:34:57.636 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:34:57.636 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:34:57.636 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@784 -- # killprocess 83961 00:34:57.636 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@954 -- # '[' -z 83961 ']' 00:34:57.636 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@958 -- # kill -0 83961 00:34:57.636 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@959 -- # uname 00:34:57.636 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:34:57.636 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 83961 00:34:57.895 killing process with pid 83961 00:34:57.895 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:34:57.895 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:34:57.895 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@972 -- # echo 'killing process with pid 83961' 00:34:57.895 Received shutdown signal, test time was about 60.000000 seconds 00:34:57.895 00:34:57.895 Latency(us) 00:34:57.895 [2024-11-20T13:55:04.954Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:34:57.895 [2024-11-20T13:55:04.954Z] =================================================================================================================== 00:34:57.895 [2024-11-20T13:55:04.954Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:34:57.895 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@973 -- # kill 83961 00:34:57.895 13:55:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@978 -- # wait 83961 00:34:57.895 [2024-11-20 13:55:04.712401] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:34:57.895 [2024-11-20 13:55:04.712524] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:34:57.895 [2024-11-20 13:55:04.712571] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:34:57.895 [2024-11-20 13:55:04.712587] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:34:57.895 [2024-11-20 13:55:04.869400] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:34:58.463 13:55:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@786 -- # return 0 00:34:58.463 00:34:58.463 real 0m16.963s 00:34:58.463 user 0m21.465s 00:34:58.463 sys 0m1.947s 00:34:58.463 ************************************ 00:34:58.463 END TEST raid_rebuild_test_sb_4k 00:34:58.463 ************************************ 00:34:58.463 13:55:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@1130 -- # xtrace_disable 00:34:58.463 13:55:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:34:58.463 13:55:05 bdev_raid -- bdev/bdev_raid.sh@1003 -- # base_malloc_params='-m 32' 00:34:58.463 13:55:05 bdev_raid -- bdev/bdev_raid.sh@1004 -- # run_test raid_state_function_test_sb_md_separate raid_state_function_test raid1 2 true 00:34:58.463 13:55:05 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:34:58.463 13:55:05 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:34:58.463 13:55:05 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:34:58.463 ************************************ 00:34:58.463 START TEST raid_state_function_test_sb_md_separate 00:34:58.463 ************************************ 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # local strip_size 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@229 -- # raid_pid=84622 00:34:58.463 Process raid pid: 84622 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 84622' 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@231 -- # waitforlisten 84622 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@835 -- # '[' -z 84622 ']' 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@840 -- # local max_retries=100 00:34:58.463 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@844 -- # xtrace_disable 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:34:58.463 13:55:05 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:34:58.722 [2024-11-20 13:55:05.579304] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:34:58.722 [2024-11-20 13:55:05.579415] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:34:58.722 [2024-11-20 13:55:05.736126] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:34:58.980 [2024-11-20 13:55:05.836315] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:34:58.980 [2024-11-20 13:55:05.958746] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:34:58.980 [2024-11-20 13:55:05.958801] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:34:59.548 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:34:59.548 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@868 -- # return 0 00:34:59.548 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:34:59.548 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:59.548 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:34:59.548 [2024-11-20 13:55:06.424161] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:34:59.548 [2024-11-20 13:55:06.424215] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:34:59.548 [2024-11-20 13:55:06.424223] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:34:59.548 [2024-11-20 13:55:06.424232] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:34:59.548 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:59.548 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:34:59.548 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:34:59.548 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:34:59.548 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:34:59.548 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:34:59.548 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:34:59.548 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:59.548 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:59.548 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:59.548 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:59.548 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:59.548 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:34:59.548 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:59.548 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:34:59.548 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:59.548 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:59.548 "name": "Existed_Raid", 00:34:59.548 "uuid": "02d63644-e2fb-4c68-8bb4-1eeba0f3e489", 00:34:59.548 "strip_size_kb": 0, 00:34:59.548 "state": "configuring", 00:34:59.548 "raid_level": "raid1", 00:34:59.548 "superblock": true, 00:34:59.548 "num_base_bdevs": 2, 00:34:59.548 "num_base_bdevs_discovered": 0, 00:34:59.548 "num_base_bdevs_operational": 2, 00:34:59.548 "base_bdevs_list": [ 00:34:59.548 { 00:34:59.548 "name": "BaseBdev1", 00:34:59.548 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:59.548 "is_configured": false, 00:34:59.548 "data_offset": 0, 00:34:59.548 "data_size": 0 00:34:59.548 }, 00:34:59.548 { 00:34:59.548 "name": "BaseBdev2", 00:34:59.548 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:59.548 "is_configured": false, 00:34:59.548 "data_offset": 0, 00:34:59.548 "data_size": 0 00:34:59.548 } 00:34:59.548 ] 00:34:59.548 }' 00:34:59.548 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:59.548 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:34:59.807 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:34:59.807 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:59.807 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:34:59.807 [2024-11-20 13:55:06.760171] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:34:59.807 [2024-11-20 13:55:06.760207] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:34:59.807 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:59.807 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:34:59.807 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:59.807 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:34:59.807 [2024-11-20 13:55:06.768160] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:34:59.807 [2024-11-20 13:55:06.768197] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:34:59.807 [2024-11-20 13:55:06.768205] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:34:59.807 [2024-11-20 13:55:06.768215] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:34:59.807 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:59.807 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev1 00:34:59.807 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:59.807 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:34:59.807 [2024-11-20 13:55:06.798991] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:34:59.807 BaseBdev1 00:34:59.807 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:59.807 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:34:59.807 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:34:59.807 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:34:59.807 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@905 -- # local i 00:34:59.807 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:34:59.807 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:34:59.807 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:34:59.807 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:59.807 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:34:59.807 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:59.807 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:34:59.807 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:59.807 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:34:59.807 [ 00:34:59.807 { 00:34:59.807 "name": "BaseBdev1", 00:34:59.807 "aliases": [ 00:34:59.807 "c18bb085-10f4-49f7-8d5c-bc26119b3034" 00:34:59.807 ], 00:34:59.807 "product_name": "Malloc disk", 00:34:59.807 "block_size": 4096, 00:34:59.807 "num_blocks": 8192, 00:34:59.807 "uuid": "c18bb085-10f4-49f7-8d5c-bc26119b3034", 00:34:59.807 "md_size": 32, 00:34:59.807 "md_interleave": false, 00:34:59.807 "dif_type": 0, 00:34:59.807 "assigned_rate_limits": { 00:34:59.807 "rw_ios_per_sec": 0, 00:34:59.807 "rw_mbytes_per_sec": 0, 00:34:59.807 "r_mbytes_per_sec": 0, 00:34:59.807 "w_mbytes_per_sec": 0 00:34:59.807 }, 00:34:59.807 "claimed": true, 00:34:59.807 "claim_type": "exclusive_write", 00:34:59.807 "zoned": false, 00:34:59.807 "supported_io_types": { 00:34:59.807 "read": true, 00:34:59.807 "write": true, 00:34:59.807 "unmap": true, 00:34:59.807 "flush": true, 00:34:59.807 "reset": true, 00:34:59.808 "nvme_admin": false, 00:34:59.808 "nvme_io": false, 00:34:59.808 "nvme_io_md": false, 00:34:59.808 "write_zeroes": true, 00:34:59.808 "zcopy": true, 00:34:59.808 "get_zone_info": false, 00:34:59.808 "zone_management": false, 00:34:59.808 "zone_append": false, 00:34:59.808 "compare": false, 00:34:59.808 "compare_and_write": false, 00:34:59.808 "abort": true, 00:34:59.808 "seek_hole": false, 00:34:59.808 "seek_data": false, 00:34:59.808 "copy": true, 00:34:59.808 "nvme_iov_md": false 00:34:59.808 }, 00:34:59.808 "memory_domains": [ 00:34:59.808 { 00:34:59.808 "dma_device_id": "system", 00:34:59.808 "dma_device_type": 1 00:34:59.808 }, 00:34:59.808 { 00:34:59.808 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:34:59.808 "dma_device_type": 2 00:34:59.808 } 00:34:59.808 ], 00:34:59.808 "driver_specific": {} 00:34:59.808 } 00:34:59.808 ] 00:34:59.808 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:59.808 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@911 -- # return 0 00:34:59.808 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:34:59.808 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:34:59.808 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:34:59.808 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:34:59.808 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:34:59.808 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:34:59.808 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:34:59.808 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:34:59.808 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:34:59.808 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:34:59.808 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:34:59.808 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:34:59.808 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:34:59.808 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:34:59.808 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:34:59.808 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:34:59.808 "name": "Existed_Raid", 00:34:59.808 "uuid": "5c8a123f-9f02-4dff-8342-519498c29d34", 00:34:59.808 "strip_size_kb": 0, 00:34:59.808 "state": "configuring", 00:34:59.808 "raid_level": "raid1", 00:34:59.808 "superblock": true, 00:34:59.808 "num_base_bdevs": 2, 00:34:59.808 "num_base_bdevs_discovered": 1, 00:34:59.808 "num_base_bdevs_operational": 2, 00:34:59.808 "base_bdevs_list": [ 00:34:59.808 { 00:34:59.808 "name": "BaseBdev1", 00:34:59.808 "uuid": "c18bb085-10f4-49f7-8d5c-bc26119b3034", 00:34:59.808 "is_configured": true, 00:34:59.808 "data_offset": 256, 00:34:59.808 "data_size": 7936 00:34:59.808 }, 00:34:59.808 { 00:34:59.808 "name": "BaseBdev2", 00:34:59.808 "uuid": "00000000-0000-0000-0000-000000000000", 00:34:59.808 "is_configured": false, 00:34:59.808 "data_offset": 0, 00:34:59.808 "data_size": 0 00:34:59.808 } 00:34:59.808 ] 00:34:59.808 }' 00:34:59.808 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:34:59.808 13:55:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:00.373 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:35:00.373 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:00.374 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:00.374 [2024-11-20 13:55:07.151122] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:35:00.374 [2024-11-20 13:55:07.151170] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:35:00.374 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:00.374 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:35:00.374 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:00.374 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:00.374 [2024-11-20 13:55:07.159149] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:35:00.374 [2024-11-20 13:55:07.160806] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:35:00.374 [2024-11-20 13:55:07.160847] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:35:00.374 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:00.374 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:35:00.374 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:35:00.374 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:35:00.374 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:35:00.374 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:35:00.374 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:00.374 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:00.374 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:35:00.374 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:00.374 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:00.374 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:00.374 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:00.374 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:00.374 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:35:00.374 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:00.374 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:00.374 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:00.374 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:00.374 "name": "Existed_Raid", 00:35:00.374 "uuid": "d6e0250b-541f-4a7a-be1d-8ef5571f4f1f", 00:35:00.374 "strip_size_kb": 0, 00:35:00.374 "state": "configuring", 00:35:00.374 "raid_level": "raid1", 00:35:00.374 "superblock": true, 00:35:00.374 "num_base_bdevs": 2, 00:35:00.374 "num_base_bdevs_discovered": 1, 00:35:00.374 "num_base_bdevs_operational": 2, 00:35:00.374 "base_bdevs_list": [ 00:35:00.374 { 00:35:00.374 "name": "BaseBdev1", 00:35:00.374 "uuid": "c18bb085-10f4-49f7-8d5c-bc26119b3034", 00:35:00.374 "is_configured": true, 00:35:00.374 "data_offset": 256, 00:35:00.374 "data_size": 7936 00:35:00.374 }, 00:35:00.374 { 00:35:00.374 "name": "BaseBdev2", 00:35:00.374 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:00.374 "is_configured": false, 00:35:00.374 "data_offset": 0, 00:35:00.374 "data_size": 0 00:35:00.374 } 00:35:00.374 ] 00:35:00.374 }' 00:35:00.374 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:00.374 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev2 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:00.633 [2024-11-20 13:55:07.519188] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:35:00.633 [2024-11-20 13:55:07.519383] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:35:00.633 [2024-11-20 13:55:07.519397] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:35:00.633 [2024-11-20 13:55:07.519469] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:35:00.633 [2024-11-20 13:55:07.519571] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:35:00.633 [2024-11-20 13:55:07.519581] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:35:00.633 [2024-11-20 13:55:07.519652] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:35:00.633 BaseBdev2 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@905 -- # local i 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:00.633 [ 00:35:00.633 { 00:35:00.633 "name": "BaseBdev2", 00:35:00.633 "aliases": [ 00:35:00.633 "7bbc37c8-9594-4a22-b80c-fdc9881437cc" 00:35:00.633 ], 00:35:00.633 "product_name": "Malloc disk", 00:35:00.633 "block_size": 4096, 00:35:00.633 "num_blocks": 8192, 00:35:00.633 "uuid": "7bbc37c8-9594-4a22-b80c-fdc9881437cc", 00:35:00.633 "md_size": 32, 00:35:00.633 "md_interleave": false, 00:35:00.633 "dif_type": 0, 00:35:00.633 "assigned_rate_limits": { 00:35:00.633 "rw_ios_per_sec": 0, 00:35:00.633 "rw_mbytes_per_sec": 0, 00:35:00.633 "r_mbytes_per_sec": 0, 00:35:00.633 "w_mbytes_per_sec": 0 00:35:00.633 }, 00:35:00.633 "claimed": true, 00:35:00.633 "claim_type": "exclusive_write", 00:35:00.633 "zoned": false, 00:35:00.633 "supported_io_types": { 00:35:00.633 "read": true, 00:35:00.633 "write": true, 00:35:00.633 "unmap": true, 00:35:00.633 "flush": true, 00:35:00.633 "reset": true, 00:35:00.633 "nvme_admin": false, 00:35:00.633 "nvme_io": false, 00:35:00.633 "nvme_io_md": false, 00:35:00.633 "write_zeroes": true, 00:35:00.633 "zcopy": true, 00:35:00.633 "get_zone_info": false, 00:35:00.633 "zone_management": false, 00:35:00.633 "zone_append": false, 00:35:00.633 "compare": false, 00:35:00.633 "compare_and_write": false, 00:35:00.633 "abort": true, 00:35:00.633 "seek_hole": false, 00:35:00.633 "seek_data": false, 00:35:00.633 "copy": true, 00:35:00.633 "nvme_iov_md": false 00:35:00.633 }, 00:35:00.633 "memory_domains": [ 00:35:00.633 { 00:35:00.633 "dma_device_id": "system", 00:35:00.633 "dma_device_type": 1 00:35:00.633 }, 00:35:00.633 { 00:35:00.633 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:35:00.633 "dma_device_type": 2 00:35:00.633 } 00:35:00.633 ], 00:35:00.633 "driver_specific": {} 00:35:00.633 } 00:35:00.633 ] 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@911 -- # return 0 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:00.633 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:00.634 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:35:00.634 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:00.634 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:00.634 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:00.634 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:00.634 "name": "Existed_Raid", 00:35:00.634 "uuid": "d6e0250b-541f-4a7a-be1d-8ef5571f4f1f", 00:35:00.634 "strip_size_kb": 0, 00:35:00.634 "state": "online", 00:35:00.634 "raid_level": "raid1", 00:35:00.634 "superblock": true, 00:35:00.634 "num_base_bdevs": 2, 00:35:00.634 "num_base_bdevs_discovered": 2, 00:35:00.634 "num_base_bdevs_operational": 2, 00:35:00.634 "base_bdevs_list": [ 00:35:00.634 { 00:35:00.634 "name": "BaseBdev1", 00:35:00.634 "uuid": "c18bb085-10f4-49f7-8d5c-bc26119b3034", 00:35:00.634 "is_configured": true, 00:35:00.634 "data_offset": 256, 00:35:00.634 "data_size": 7936 00:35:00.634 }, 00:35:00.634 { 00:35:00.634 "name": "BaseBdev2", 00:35:00.634 "uuid": "7bbc37c8-9594-4a22-b80c-fdc9881437cc", 00:35:00.634 "is_configured": true, 00:35:00.634 "data_offset": 256, 00:35:00.634 "data_size": 7936 00:35:00.634 } 00:35:00.634 ] 00:35:00.634 }' 00:35:00.634 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:00.634 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:00.892 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:35:00.892 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:35:00.892 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:35:00.892 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:35:00.892 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:35:00.892 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:35:00.892 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:35:00.892 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:35:00.892 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:00.892 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:00.892 [2024-11-20 13:55:07.879615] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:35:00.892 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:00.892 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:35:00.892 "name": "Existed_Raid", 00:35:00.892 "aliases": [ 00:35:00.892 "d6e0250b-541f-4a7a-be1d-8ef5571f4f1f" 00:35:00.892 ], 00:35:00.892 "product_name": "Raid Volume", 00:35:00.892 "block_size": 4096, 00:35:00.892 "num_blocks": 7936, 00:35:00.892 "uuid": "d6e0250b-541f-4a7a-be1d-8ef5571f4f1f", 00:35:00.892 "md_size": 32, 00:35:00.892 "md_interleave": false, 00:35:00.892 "dif_type": 0, 00:35:00.892 "assigned_rate_limits": { 00:35:00.892 "rw_ios_per_sec": 0, 00:35:00.892 "rw_mbytes_per_sec": 0, 00:35:00.892 "r_mbytes_per_sec": 0, 00:35:00.892 "w_mbytes_per_sec": 0 00:35:00.892 }, 00:35:00.892 "claimed": false, 00:35:00.892 "zoned": false, 00:35:00.892 "supported_io_types": { 00:35:00.892 "read": true, 00:35:00.892 "write": true, 00:35:00.892 "unmap": false, 00:35:00.892 "flush": false, 00:35:00.892 "reset": true, 00:35:00.892 "nvme_admin": false, 00:35:00.892 "nvme_io": false, 00:35:00.892 "nvme_io_md": false, 00:35:00.892 "write_zeroes": true, 00:35:00.892 "zcopy": false, 00:35:00.892 "get_zone_info": false, 00:35:00.892 "zone_management": false, 00:35:00.892 "zone_append": false, 00:35:00.892 "compare": false, 00:35:00.892 "compare_and_write": false, 00:35:00.892 "abort": false, 00:35:00.892 "seek_hole": false, 00:35:00.892 "seek_data": false, 00:35:00.892 "copy": false, 00:35:00.892 "nvme_iov_md": false 00:35:00.892 }, 00:35:00.892 "memory_domains": [ 00:35:00.892 { 00:35:00.892 "dma_device_id": "system", 00:35:00.892 "dma_device_type": 1 00:35:00.892 }, 00:35:00.892 { 00:35:00.892 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:35:00.892 "dma_device_type": 2 00:35:00.892 }, 00:35:00.892 { 00:35:00.892 "dma_device_id": "system", 00:35:00.892 "dma_device_type": 1 00:35:00.892 }, 00:35:00.892 { 00:35:00.892 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:35:00.892 "dma_device_type": 2 00:35:00.892 } 00:35:00.892 ], 00:35:00.892 "driver_specific": { 00:35:00.892 "raid": { 00:35:00.892 "uuid": "d6e0250b-541f-4a7a-be1d-8ef5571f4f1f", 00:35:00.892 "strip_size_kb": 0, 00:35:00.892 "state": "online", 00:35:00.892 "raid_level": "raid1", 00:35:00.892 "superblock": true, 00:35:00.892 "num_base_bdevs": 2, 00:35:00.892 "num_base_bdevs_discovered": 2, 00:35:00.892 "num_base_bdevs_operational": 2, 00:35:00.892 "base_bdevs_list": [ 00:35:00.892 { 00:35:00.892 "name": "BaseBdev1", 00:35:00.892 "uuid": "c18bb085-10f4-49f7-8d5c-bc26119b3034", 00:35:00.892 "is_configured": true, 00:35:00.892 "data_offset": 256, 00:35:00.892 "data_size": 7936 00:35:00.892 }, 00:35:00.892 { 00:35:00.892 "name": "BaseBdev2", 00:35:00.892 "uuid": "7bbc37c8-9594-4a22-b80c-fdc9881437cc", 00:35:00.892 "is_configured": true, 00:35:00.892 "data_offset": 256, 00:35:00.892 "data_size": 7936 00:35:00.892 } 00:35:00.892 ] 00:35:00.892 } 00:35:00.892 } 00:35:00.892 }' 00:35:00.892 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:35:00.892 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:35:00.892 BaseBdev2' 00:35:00.893 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:35:01.152 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:35:01.152 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:35:01.152 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:35:01.152 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:35:01.152 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:01.152 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:01.152 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:01.152 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:35:01.152 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:35:01.152 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:35:01.152 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:35:01.152 13:55:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:35:01.152 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:01.152 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:01.152 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:01.152 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:35:01.152 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:35:01.152 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:35:01.152 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:01.152 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:01.152 [2024-11-20 13:55:08.039388] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:35:01.152 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:01.152 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@260 -- # local expected_state 00:35:01.152 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:35:01.152 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@198 -- # case $1 in 00:35:01.152 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@199 -- # return 0 00:35:01.152 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:35:01.153 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:35:01.153 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:35:01.153 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:01.153 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:01.153 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:01.153 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:35:01.153 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:01.153 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:01.153 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:01.153 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:01.153 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:01.153 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:35:01.153 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:01.153 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:01.153 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:01.153 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:01.153 "name": "Existed_Raid", 00:35:01.153 "uuid": "d6e0250b-541f-4a7a-be1d-8ef5571f4f1f", 00:35:01.153 "strip_size_kb": 0, 00:35:01.153 "state": "online", 00:35:01.153 "raid_level": "raid1", 00:35:01.153 "superblock": true, 00:35:01.153 "num_base_bdevs": 2, 00:35:01.153 "num_base_bdevs_discovered": 1, 00:35:01.153 "num_base_bdevs_operational": 1, 00:35:01.153 "base_bdevs_list": [ 00:35:01.153 { 00:35:01.153 "name": null, 00:35:01.153 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:01.153 "is_configured": false, 00:35:01.153 "data_offset": 0, 00:35:01.153 "data_size": 7936 00:35:01.153 }, 00:35:01.153 { 00:35:01.153 "name": "BaseBdev2", 00:35:01.153 "uuid": "7bbc37c8-9594-4a22-b80c-fdc9881437cc", 00:35:01.153 "is_configured": true, 00:35:01.153 "data_offset": 256, 00:35:01.153 "data_size": 7936 00:35:01.153 } 00:35:01.153 ] 00:35:01.153 }' 00:35:01.153 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:01.153 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:01.412 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:35:01.412 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:35:01.412 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:35:01.412 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:01.412 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:01.412 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:01.412 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:01.412 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:35:01.412 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:35:01.412 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:35:01.412 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:01.412 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:01.412 [2024-11-20 13:55:08.440890] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:35:01.412 [2024-11-20 13:55:08.440996] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:35:01.671 [2024-11-20 13:55:08.494236] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:35:01.672 [2024-11-20 13:55:08.494405] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:35:01.672 [2024-11-20 13:55:08.494472] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:35:01.672 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:01.672 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:35:01.672 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:35:01.672 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:01.672 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:35:01.672 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:01.672 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:01.672 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:01.672 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:35:01.672 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:35:01.672 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:35:01.672 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@326 -- # killprocess 84622 00:35:01.672 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@954 -- # '[' -z 84622 ']' 00:35:01.672 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@958 -- # kill -0 84622 00:35:01.672 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@959 -- # uname 00:35:01.672 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:35:01.672 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 84622 00:35:01.672 killing process with pid 84622 00:35:01.672 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:35:01.672 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:35:01.672 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@972 -- # echo 'killing process with pid 84622' 00:35:01.672 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@973 -- # kill 84622 00:35:01.672 [2024-11-20 13:55:08.556643] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:35:01.672 13:55:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@978 -- # wait 84622 00:35:01.672 [2024-11-20 13:55:08.565623] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:35:02.239 13:55:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@328 -- # return 0 00:35:02.239 00:35:02.239 real 0m3.672s 00:35:02.239 user 0m5.317s 00:35:02.239 sys 0m0.602s 00:35:02.239 13:55:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@1130 -- # xtrace_disable 00:35:02.239 13:55:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:02.239 ************************************ 00:35:02.239 END TEST raid_state_function_test_sb_md_separate 00:35:02.239 ************************************ 00:35:02.239 13:55:09 bdev_raid -- bdev/bdev_raid.sh@1005 -- # run_test raid_superblock_test_md_separate raid_superblock_test raid1 2 00:35:02.239 13:55:09 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:35:02.239 13:55:09 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:35:02.239 13:55:09 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:35:02.239 ************************************ 00:35:02.239 START TEST raid_superblock_test_md_separate 00:35:02.239 ************************************ 00:35:02.239 13:55:09 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:35:02.239 13:55:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:35:02.239 13:55:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:35:02.239 13:55:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:35:02.239 13:55:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:35:02.239 13:55:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:35:02.239 13:55:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:35:02.239 13:55:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:35:02.239 13:55:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:35:02.239 13:55:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:35:02.239 13:55:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@399 -- # local strip_size 00:35:02.239 13:55:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:35:02.239 13:55:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:35:02.239 13:55:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:35:02.239 13:55:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:35:02.239 13:55:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:35:02.239 13:55:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@412 -- # raid_pid=84859 00:35:02.239 13:55:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@413 -- # waitforlisten 84859 00:35:02.239 13:55:09 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@835 -- # '[' -z 84859 ']' 00:35:02.239 13:55:09 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:35:02.239 13:55:09 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@840 -- # local max_retries=100 00:35:02.239 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:35:02.239 13:55:09 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:35:02.239 13:55:09 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@844 -- # xtrace_disable 00:35:02.239 13:55:09 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:02.239 13:55:09 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:35:02.497 [2024-11-20 13:55:09.309719] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:35:02.497 [2024-11-20 13:55:09.309856] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid84859 ] 00:35:02.497 [2024-11-20 13:55:09.467272] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:35:02.755 [2024-11-20 13:55:09.567399] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:35:02.755 [2024-11-20 13:55:09.687063] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:35:02.755 [2024-11-20 13:55:09.687102] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:35:03.323 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:35:03.323 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@868 -- # return 0 00:35:03.323 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:35:03.323 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:35:03.323 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:35:03.323 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:35:03.323 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:35:03.323 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:35:03.323 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:35:03.323 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b malloc1 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:03.324 malloc1 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:03.324 [2024-11-20 13:55:10.238441] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:35:03.324 [2024-11-20 13:55:10.238599] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:35:03.324 [2024-11-20 13:55:10.238626] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:35:03.324 [2024-11-20 13:55:10.238635] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:35:03.324 [2024-11-20 13:55:10.240361] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:35:03.324 [2024-11-20 13:55:10.240390] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:35:03.324 pt1 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b malloc2 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:03.324 malloc2 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:03.324 [2024-11-20 13:55:10.272649] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:35:03.324 [2024-11-20 13:55:10.272693] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:35:03.324 [2024-11-20 13:55:10.272710] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:35:03.324 [2024-11-20 13:55:10.272718] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:35:03.324 [2024-11-20 13:55:10.274363] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:35:03.324 [2024-11-20 13:55:10.274391] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:35:03.324 pt2 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:03.324 [2024-11-20 13:55:10.280672] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:35:03.324 [2024-11-20 13:55:10.282343] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:35:03.324 [2024-11-20 13:55:10.282552] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:35:03.324 [2024-11-20 13:55:10.282623] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:35:03.324 [2024-11-20 13:55:10.282705] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:35:03.324 [2024-11-20 13:55:10.282923] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:35:03.324 [2024-11-20 13:55:10.283007] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:35:03.324 [2024-11-20 13:55:10.283156] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:03.324 "name": "raid_bdev1", 00:35:03.324 "uuid": "dd1419e7-cbc8-44a7-adc9-8e48aa13b5f0", 00:35:03.324 "strip_size_kb": 0, 00:35:03.324 "state": "online", 00:35:03.324 "raid_level": "raid1", 00:35:03.324 "superblock": true, 00:35:03.324 "num_base_bdevs": 2, 00:35:03.324 "num_base_bdevs_discovered": 2, 00:35:03.324 "num_base_bdevs_operational": 2, 00:35:03.324 "base_bdevs_list": [ 00:35:03.324 { 00:35:03.324 "name": "pt1", 00:35:03.324 "uuid": "00000000-0000-0000-0000-000000000001", 00:35:03.324 "is_configured": true, 00:35:03.324 "data_offset": 256, 00:35:03.324 "data_size": 7936 00:35:03.324 }, 00:35:03.324 { 00:35:03.324 "name": "pt2", 00:35:03.324 "uuid": "00000000-0000-0000-0000-000000000002", 00:35:03.324 "is_configured": true, 00:35:03.324 "data_offset": 256, 00:35:03.324 "data_size": 7936 00:35:03.324 } 00:35:03.324 ] 00:35:03.324 }' 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:03.324 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:03.583 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:35:03.583 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:35:03.583 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:35:03.583 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:35:03.583 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:35:03.583 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:35:03.583 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:35:03.583 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:03.583 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:03.583 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:35:03.583 [2024-11-20 13:55:10.613037] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:35:03.583 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:03.583 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:35:03.583 "name": "raid_bdev1", 00:35:03.583 "aliases": [ 00:35:03.583 "dd1419e7-cbc8-44a7-adc9-8e48aa13b5f0" 00:35:03.583 ], 00:35:03.583 "product_name": "Raid Volume", 00:35:03.583 "block_size": 4096, 00:35:03.583 "num_blocks": 7936, 00:35:03.583 "uuid": "dd1419e7-cbc8-44a7-adc9-8e48aa13b5f0", 00:35:03.583 "md_size": 32, 00:35:03.583 "md_interleave": false, 00:35:03.583 "dif_type": 0, 00:35:03.583 "assigned_rate_limits": { 00:35:03.583 "rw_ios_per_sec": 0, 00:35:03.583 "rw_mbytes_per_sec": 0, 00:35:03.583 "r_mbytes_per_sec": 0, 00:35:03.583 "w_mbytes_per_sec": 0 00:35:03.583 }, 00:35:03.583 "claimed": false, 00:35:03.583 "zoned": false, 00:35:03.583 "supported_io_types": { 00:35:03.583 "read": true, 00:35:03.583 "write": true, 00:35:03.583 "unmap": false, 00:35:03.583 "flush": false, 00:35:03.583 "reset": true, 00:35:03.583 "nvme_admin": false, 00:35:03.583 "nvme_io": false, 00:35:03.583 "nvme_io_md": false, 00:35:03.583 "write_zeroes": true, 00:35:03.583 "zcopy": false, 00:35:03.583 "get_zone_info": false, 00:35:03.583 "zone_management": false, 00:35:03.583 "zone_append": false, 00:35:03.583 "compare": false, 00:35:03.583 "compare_and_write": false, 00:35:03.583 "abort": false, 00:35:03.583 "seek_hole": false, 00:35:03.583 "seek_data": false, 00:35:03.583 "copy": false, 00:35:03.583 "nvme_iov_md": false 00:35:03.583 }, 00:35:03.583 "memory_domains": [ 00:35:03.583 { 00:35:03.583 "dma_device_id": "system", 00:35:03.583 "dma_device_type": 1 00:35:03.583 }, 00:35:03.583 { 00:35:03.583 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:35:03.583 "dma_device_type": 2 00:35:03.583 }, 00:35:03.583 { 00:35:03.583 "dma_device_id": "system", 00:35:03.583 "dma_device_type": 1 00:35:03.583 }, 00:35:03.583 { 00:35:03.583 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:35:03.583 "dma_device_type": 2 00:35:03.583 } 00:35:03.583 ], 00:35:03.583 "driver_specific": { 00:35:03.583 "raid": { 00:35:03.583 "uuid": "dd1419e7-cbc8-44a7-adc9-8e48aa13b5f0", 00:35:03.583 "strip_size_kb": 0, 00:35:03.583 "state": "online", 00:35:03.583 "raid_level": "raid1", 00:35:03.583 "superblock": true, 00:35:03.583 "num_base_bdevs": 2, 00:35:03.583 "num_base_bdevs_discovered": 2, 00:35:03.583 "num_base_bdevs_operational": 2, 00:35:03.583 "base_bdevs_list": [ 00:35:03.583 { 00:35:03.583 "name": "pt1", 00:35:03.583 "uuid": "00000000-0000-0000-0000-000000000001", 00:35:03.583 "is_configured": true, 00:35:03.583 "data_offset": 256, 00:35:03.583 "data_size": 7936 00:35:03.583 }, 00:35:03.583 { 00:35:03.583 "name": "pt2", 00:35:03.583 "uuid": "00000000-0000-0000-0000-000000000002", 00:35:03.583 "is_configured": true, 00:35:03.583 "data_offset": 256, 00:35:03.583 "data_size": 7936 00:35:03.583 } 00:35:03.583 ] 00:35:03.583 } 00:35:03.583 } 00:35:03.583 }' 00:35:03.583 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:35:03.843 pt2' 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:03.843 [2024-11-20 13:55:10.793005] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=dd1419e7-cbc8-44a7-adc9-8e48aa13b5f0 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@436 -- # '[' -z dd1419e7-cbc8-44a7-adc9-8e48aa13b5f0 ']' 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:03.843 [2024-11-20 13:55:10.824742] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:35:03.843 [2024-11-20 13:55:10.824833] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:35:03.843 [2024-11-20 13:55:10.824927] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:35:03.843 [2024-11-20 13:55:10.825000] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:35:03.843 [2024-11-20 13:55:10.825012] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:35:03.843 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@652 -- # local es=0 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:04.103 [2024-11-20 13:55:10.912799] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:35:04.103 [2024-11-20 13:55:10.914509] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:35:04.103 [2024-11-20 13:55:10.914571] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:35:04.103 [2024-11-20 13:55:10.914619] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:35:04.103 [2024-11-20 13:55:10.914632] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:35:04.103 [2024-11-20 13:55:10.914642] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:35:04.103 request: 00:35:04.103 { 00:35:04.103 "name": "raid_bdev1", 00:35:04.103 "raid_level": "raid1", 00:35:04.103 "base_bdevs": [ 00:35:04.103 "malloc1", 00:35:04.103 "malloc2" 00:35:04.103 ], 00:35:04.103 "superblock": false, 00:35:04.103 "method": "bdev_raid_create", 00:35:04.103 "req_id": 1 00:35:04.103 } 00:35:04.103 Got JSON-RPC error response 00:35:04.103 response: 00:35:04.103 { 00:35:04.103 "code": -17, 00:35:04.103 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:35:04.103 } 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@655 -- # es=1 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:04.103 [2024-11-20 13:55:10.956760] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:35:04.103 [2024-11-20 13:55:10.956806] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:35:04.103 [2024-11-20 13:55:10.956820] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:35:04.103 [2024-11-20 13:55:10.956829] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:35:04.103 [2024-11-20 13:55:10.958591] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:35:04.103 [2024-11-20 13:55:10.958621] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:35:04.103 [2024-11-20 13:55:10.958663] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:35:04.103 [2024-11-20 13:55:10.958710] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:35:04.103 pt1 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:04.103 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:04.104 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:04.104 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:04.104 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:04.104 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:04.104 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:04.104 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:04.104 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:04.104 "name": "raid_bdev1", 00:35:04.104 "uuid": "dd1419e7-cbc8-44a7-adc9-8e48aa13b5f0", 00:35:04.104 "strip_size_kb": 0, 00:35:04.104 "state": "configuring", 00:35:04.104 "raid_level": "raid1", 00:35:04.104 "superblock": true, 00:35:04.104 "num_base_bdevs": 2, 00:35:04.104 "num_base_bdevs_discovered": 1, 00:35:04.104 "num_base_bdevs_operational": 2, 00:35:04.104 "base_bdevs_list": [ 00:35:04.104 { 00:35:04.104 "name": "pt1", 00:35:04.104 "uuid": "00000000-0000-0000-0000-000000000001", 00:35:04.104 "is_configured": true, 00:35:04.104 "data_offset": 256, 00:35:04.104 "data_size": 7936 00:35:04.104 }, 00:35:04.104 { 00:35:04.104 "name": null, 00:35:04.104 "uuid": "00000000-0000-0000-0000-000000000002", 00:35:04.104 "is_configured": false, 00:35:04.104 "data_offset": 256, 00:35:04.104 "data_size": 7936 00:35:04.104 } 00:35:04.104 ] 00:35:04.104 }' 00:35:04.104 13:55:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:04.104 13:55:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:04.363 [2024-11-20 13:55:11.260861] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:35:04.363 [2024-11-20 13:55:11.260931] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:35:04.363 [2024-11-20 13:55:11.260949] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:35:04.363 [2024-11-20 13:55:11.260959] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:35:04.363 [2024-11-20 13:55:11.261180] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:35:04.363 [2024-11-20 13:55:11.261194] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:35:04.363 [2024-11-20 13:55:11.261238] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:35:04.363 [2024-11-20 13:55:11.261258] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:35:04.363 [2024-11-20 13:55:11.261351] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:35:04.363 [2024-11-20 13:55:11.261360] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:35:04.363 [2024-11-20 13:55:11.261422] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:35:04.363 [2024-11-20 13:55:11.261511] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:35:04.363 [2024-11-20 13:55:11.261517] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:35:04.363 [2024-11-20 13:55:11.261596] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:35:04.363 pt2 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:04.363 "name": "raid_bdev1", 00:35:04.363 "uuid": "dd1419e7-cbc8-44a7-adc9-8e48aa13b5f0", 00:35:04.363 "strip_size_kb": 0, 00:35:04.363 "state": "online", 00:35:04.363 "raid_level": "raid1", 00:35:04.363 "superblock": true, 00:35:04.363 "num_base_bdevs": 2, 00:35:04.363 "num_base_bdevs_discovered": 2, 00:35:04.363 "num_base_bdevs_operational": 2, 00:35:04.363 "base_bdevs_list": [ 00:35:04.363 { 00:35:04.363 "name": "pt1", 00:35:04.363 "uuid": "00000000-0000-0000-0000-000000000001", 00:35:04.363 "is_configured": true, 00:35:04.363 "data_offset": 256, 00:35:04.363 "data_size": 7936 00:35:04.363 }, 00:35:04.363 { 00:35:04.363 "name": "pt2", 00:35:04.363 "uuid": "00000000-0000-0000-0000-000000000002", 00:35:04.363 "is_configured": true, 00:35:04.363 "data_offset": 256, 00:35:04.363 "data_size": 7936 00:35:04.363 } 00:35:04.363 ] 00:35:04.363 }' 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:04.363 13:55:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:04.622 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:35:04.622 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:35:04.622 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:35:04.622 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:35:04.622 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:35:04.622 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:35:04.622 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:35:04.622 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:35:04.622 13:55:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:04.622 13:55:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:04.622 [2024-11-20 13:55:11.585193] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:35:04.622 13:55:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:04.622 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:35:04.622 "name": "raid_bdev1", 00:35:04.622 "aliases": [ 00:35:04.622 "dd1419e7-cbc8-44a7-adc9-8e48aa13b5f0" 00:35:04.622 ], 00:35:04.622 "product_name": "Raid Volume", 00:35:04.622 "block_size": 4096, 00:35:04.622 "num_blocks": 7936, 00:35:04.622 "uuid": "dd1419e7-cbc8-44a7-adc9-8e48aa13b5f0", 00:35:04.622 "md_size": 32, 00:35:04.622 "md_interleave": false, 00:35:04.622 "dif_type": 0, 00:35:04.622 "assigned_rate_limits": { 00:35:04.622 "rw_ios_per_sec": 0, 00:35:04.622 "rw_mbytes_per_sec": 0, 00:35:04.622 "r_mbytes_per_sec": 0, 00:35:04.622 "w_mbytes_per_sec": 0 00:35:04.622 }, 00:35:04.622 "claimed": false, 00:35:04.622 "zoned": false, 00:35:04.622 "supported_io_types": { 00:35:04.622 "read": true, 00:35:04.622 "write": true, 00:35:04.622 "unmap": false, 00:35:04.622 "flush": false, 00:35:04.622 "reset": true, 00:35:04.622 "nvme_admin": false, 00:35:04.622 "nvme_io": false, 00:35:04.622 "nvme_io_md": false, 00:35:04.622 "write_zeroes": true, 00:35:04.622 "zcopy": false, 00:35:04.622 "get_zone_info": false, 00:35:04.622 "zone_management": false, 00:35:04.622 "zone_append": false, 00:35:04.622 "compare": false, 00:35:04.622 "compare_and_write": false, 00:35:04.622 "abort": false, 00:35:04.622 "seek_hole": false, 00:35:04.622 "seek_data": false, 00:35:04.622 "copy": false, 00:35:04.622 "nvme_iov_md": false 00:35:04.622 }, 00:35:04.622 "memory_domains": [ 00:35:04.622 { 00:35:04.622 "dma_device_id": "system", 00:35:04.622 "dma_device_type": 1 00:35:04.622 }, 00:35:04.622 { 00:35:04.622 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:35:04.622 "dma_device_type": 2 00:35:04.622 }, 00:35:04.622 { 00:35:04.622 "dma_device_id": "system", 00:35:04.622 "dma_device_type": 1 00:35:04.622 }, 00:35:04.622 { 00:35:04.622 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:35:04.622 "dma_device_type": 2 00:35:04.622 } 00:35:04.622 ], 00:35:04.622 "driver_specific": { 00:35:04.622 "raid": { 00:35:04.622 "uuid": "dd1419e7-cbc8-44a7-adc9-8e48aa13b5f0", 00:35:04.622 "strip_size_kb": 0, 00:35:04.622 "state": "online", 00:35:04.622 "raid_level": "raid1", 00:35:04.622 "superblock": true, 00:35:04.622 "num_base_bdevs": 2, 00:35:04.622 "num_base_bdevs_discovered": 2, 00:35:04.622 "num_base_bdevs_operational": 2, 00:35:04.622 "base_bdevs_list": [ 00:35:04.622 { 00:35:04.622 "name": "pt1", 00:35:04.622 "uuid": "00000000-0000-0000-0000-000000000001", 00:35:04.622 "is_configured": true, 00:35:04.622 "data_offset": 256, 00:35:04.622 "data_size": 7936 00:35:04.622 }, 00:35:04.622 { 00:35:04.622 "name": "pt2", 00:35:04.622 "uuid": "00000000-0000-0000-0000-000000000002", 00:35:04.622 "is_configured": true, 00:35:04.622 "data_offset": 256, 00:35:04.622 "data_size": 7936 00:35:04.622 } 00:35:04.622 ] 00:35:04.622 } 00:35:04.622 } 00:35:04.622 }' 00:35:04.622 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:35:04.622 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:35:04.622 pt2' 00:35:04.622 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:35:04.622 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:35:04.622 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:35:04.622 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:35:04.622 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:35:04.622 13:55:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:04.622 13:55:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:04.880 [2024-11-20 13:55:11.749197] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # '[' dd1419e7-cbc8-44a7-adc9-8e48aa13b5f0 '!=' dd1419e7-cbc8-44a7-adc9-8e48aa13b5f0 ']' 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@198 -- # case $1 in 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@199 -- # return 0 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:04.880 [2024-11-20 13:55:11.776975] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:04.880 "name": "raid_bdev1", 00:35:04.880 "uuid": "dd1419e7-cbc8-44a7-adc9-8e48aa13b5f0", 00:35:04.880 "strip_size_kb": 0, 00:35:04.880 "state": "online", 00:35:04.880 "raid_level": "raid1", 00:35:04.880 "superblock": true, 00:35:04.880 "num_base_bdevs": 2, 00:35:04.880 "num_base_bdevs_discovered": 1, 00:35:04.880 "num_base_bdevs_operational": 1, 00:35:04.880 "base_bdevs_list": [ 00:35:04.880 { 00:35:04.880 "name": null, 00:35:04.880 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:04.880 "is_configured": false, 00:35:04.880 "data_offset": 0, 00:35:04.880 "data_size": 7936 00:35:04.880 }, 00:35:04.880 { 00:35:04.880 "name": "pt2", 00:35:04.880 "uuid": "00000000-0000-0000-0000-000000000002", 00:35:04.880 "is_configured": true, 00:35:04.880 "data_offset": 256, 00:35:04.880 "data_size": 7936 00:35:04.880 } 00:35:04.880 ] 00:35:04.880 }' 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:04.880 13:55:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:05.139 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:35:05.139 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:05.139 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:05.139 [2024-11-20 13:55:12.117034] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:35:05.139 [2024-11-20 13:55:12.117063] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:35:05.139 [2024-11-20 13:55:12.117134] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:35:05.139 [2024-11-20 13:55:12.117181] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:35:05.139 [2024-11-20 13:55:12.117191] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:35:05.139 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:05.139 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:05.139 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:05.139 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:05.139 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:35:05.139 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:05.139 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:35:05.139 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:35:05.139 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:35:05.139 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:35:05.139 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:35:05.139 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:05.139 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:05.139 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:05.139 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:35:05.139 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:35:05.139 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:35:05.139 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:35:05.139 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@519 -- # i=1 00:35:05.139 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:35:05.139 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:05.139 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:05.139 [2024-11-20 13:55:12.177027] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:35:05.139 [2024-11-20 13:55:12.177076] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:35:05.139 [2024-11-20 13:55:12.177090] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:35:05.140 [2024-11-20 13:55:12.177099] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:35:05.140 [2024-11-20 13:55:12.178883] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:35:05.140 [2024-11-20 13:55:12.178915] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:35:05.140 [2024-11-20 13:55:12.178959] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:35:05.140 [2024-11-20 13:55:12.179013] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:35:05.140 [2024-11-20 13:55:12.179101] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:35:05.140 [2024-11-20 13:55:12.179112] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:35:05.140 [2024-11-20 13:55:12.179177] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:35:05.140 [2024-11-20 13:55:12.179261] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:35:05.140 [2024-11-20 13:55:12.179267] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:35:05.140 [2024-11-20 13:55:12.179345] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:35:05.140 pt2 00:35:05.140 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:05.140 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:35:05.140 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:05.140 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:05.140 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:05.140 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:05.140 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:35:05.140 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:05.140 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:05.140 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:05.140 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:05.140 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:05.140 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:05.140 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:05.140 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:05.140 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:05.398 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:05.398 "name": "raid_bdev1", 00:35:05.398 "uuid": "dd1419e7-cbc8-44a7-adc9-8e48aa13b5f0", 00:35:05.398 "strip_size_kb": 0, 00:35:05.398 "state": "online", 00:35:05.398 "raid_level": "raid1", 00:35:05.398 "superblock": true, 00:35:05.398 "num_base_bdevs": 2, 00:35:05.398 "num_base_bdevs_discovered": 1, 00:35:05.398 "num_base_bdevs_operational": 1, 00:35:05.398 "base_bdevs_list": [ 00:35:05.398 { 00:35:05.398 "name": null, 00:35:05.398 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:05.398 "is_configured": false, 00:35:05.398 "data_offset": 256, 00:35:05.398 "data_size": 7936 00:35:05.398 }, 00:35:05.398 { 00:35:05.398 "name": "pt2", 00:35:05.398 "uuid": "00000000-0000-0000-0000-000000000002", 00:35:05.398 "is_configured": true, 00:35:05.398 "data_offset": 256, 00:35:05.398 "data_size": 7936 00:35:05.398 } 00:35:05.398 ] 00:35:05.398 }' 00:35:05.398 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:05.398 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:05.655 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:35:05.655 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:05.655 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:05.656 [2024-11-20 13:55:12.497073] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:35:05.656 [2024-11-20 13:55:12.497099] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:35:05.656 [2024-11-20 13:55:12.497167] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:35:05.656 [2024-11-20 13:55:12.497216] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:35:05.656 [2024-11-20 13:55:12.497224] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:05.656 [2024-11-20 13:55:12.541100] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:35:05.656 [2024-11-20 13:55:12.541155] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:35:05.656 [2024-11-20 13:55:12.541173] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:35:05.656 [2024-11-20 13:55:12.541180] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:35:05.656 [2024-11-20 13:55:12.543000] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:35:05.656 [2024-11-20 13:55:12.543034] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:35:05.656 [2024-11-20 13:55:12.543083] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:35:05.656 [2024-11-20 13:55:12.543120] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:35:05.656 [2024-11-20 13:55:12.543223] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:35:05.656 [2024-11-20 13:55:12.543232] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:35:05.656 [2024-11-20 13:55:12.543249] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:35:05.656 [2024-11-20 13:55:12.543297] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:35:05.656 [2024-11-20 13:55:12.543357] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:35:05.656 [2024-11-20 13:55:12.543364] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:35:05.656 [2024-11-20 13:55:12.543424] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:35:05.656 [2024-11-20 13:55:12.543505] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:35:05.656 [2024-11-20 13:55:12.543518] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:35:05.656 [2024-11-20 13:55:12.543599] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:35:05.656 pt1 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:05.656 "name": "raid_bdev1", 00:35:05.656 "uuid": "dd1419e7-cbc8-44a7-adc9-8e48aa13b5f0", 00:35:05.656 "strip_size_kb": 0, 00:35:05.656 "state": "online", 00:35:05.656 "raid_level": "raid1", 00:35:05.656 "superblock": true, 00:35:05.656 "num_base_bdevs": 2, 00:35:05.656 "num_base_bdevs_discovered": 1, 00:35:05.656 "num_base_bdevs_operational": 1, 00:35:05.656 "base_bdevs_list": [ 00:35:05.656 { 00:35:05.656 "name": null, 00:35:05.656 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:05.656 "is_configured": false, 00:35:05.656 "data_offset": 256, 00:35:05.656 "data_size": 7936 00:35:05.656 }, 00:35:05.656 { 00:35:05.656 "name": "pt2", 00:35:05.656 "uuid": "00000000-0000-0000-0000-000000000002", 00:35:05.656 "is_configured": true, 00:35:05.656 "data_offset": 256, 00:35:05.656 "data_size": 7936 00:35:05.656 } 00:35:05.656 ] 00:35:05.656 }' 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:05.656 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:05.914 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:35:05.914 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:05.914 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:05.914 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:35:05.914 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:05.914 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:35:05.914 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:35:05.914 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:05.914 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:35:05.914 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:05.914 [2024-11-20 13:55:12.889375] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:35:05.914 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:05.914 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # '[' dd1419e7-cbc8-44a7-adc9-8e48aa13b5f0 '!=' dd1419e7-cbc8-44a7-adc9-8e48aa13b5f0 ']' 00:35:05.914 13:55:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@563 -- # killprocess 84859 00:35:05.914 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@954 -- # '[' -z 84859 ']' 00:35:05.914 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@958 -- # kill -0 84859 00:35:05.914 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@959 -- # uname 00:35:05.914 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:35:05.914 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 84859 00:35:05.914 killing process with pid 84859 00:35:05.914 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:35:05.914 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:35:05.914 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@972 -- # echo 'killing process with pid 84859' 00:35:05.914 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@973 -- # kill 84859 00:35:05.914 [2024-11-20 13:55:12.938646] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:35:05.914 13:55:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@978 -- # wait 84859 00:35:05.914 [2024-11-20 13:55:12.938732] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:35:05.914 [2024-11-20 13:55:12.938778] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:35:05.914 [2024-11-20 13:55:12.938793] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:35:06.184 [2024-11-20 13:55:13.052982] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:35:06.776 13:55:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@565 -- # return 0 00:35:06.776 00:35:06.776 real 0m4.416s 00:35:06.776 user 0m6.721s 00:35:06.776 sys 0m0.789s 00:35:06.776 13:55:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@1130 -- # xtrace_disable 00:35:06.776 13:55:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:06.776 ************************************ 00:35:06.776 END TEST raid_superblock_test_md_separate 00:35:06.776 ************************************ 00:35:06.776 13:55:13 bdev_raid -- bdev/bdev_raid.sh@1006 -- # '[' true = true ']' 00:35:06.776 13:55:13 bdev_raid -- bdev/bdev_raid.sh@1007 -- # run_test raid_rebuild_test_sb_md_separate raid_rebuild_test raid1 2 true false true 00:35:06.776 13:55:13 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:35:06.776 13:55:13 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:35:06.776 13:55:13 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:35:06.776 ************************************ 00:35:06.776 START TEST raid_rebuild_test_sb_md_separate 00:35:06.776 ************************************ 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false true 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@573 -- # local verify=true 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # local strip_size 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@577 -- # local create_arg 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@579 -- # local data_offset 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@597 -- # raid_pid=85165 00:35:06.776 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@598 -- # waitforlisten 85165 00:35:06.776 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:35:06.777 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@835 -- # '[' -z 85165 ']' 00:35:06.777 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:35:06.777 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@840 -- # local max_retries=100 00:35:06.777 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:35:06.777 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@844 -- # xtrace_disable 00:35:06.777 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:06.777 13:55:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:35:06.777 I/O size of 3145728 is greater than zero copy threshold (65536). 00:35:06.777 Zero copy mechanism will not be used. 00:35:06.777 [2024-11-20 13:55:13.770199] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:35:06.777 [2024-11-20 13:55:13.770295] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85165 ] 00:35:07.035 [2024-11-20 13:55:13.924624] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:35:07.035 [2024-11-20 13:55:14.056867] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:35:07.293 [2024-11-20 13:55:14.207968] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:35:07.293 [2024-11-20 13:55:14.208039] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:35:07.551 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:35:07.551 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@868 -- # return 0 00:35:07.551 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:35:07.551 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev1_malloc 00:35:07.551 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:07.551 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:07.551 BaseBdev1_malloc 00:35:07.551 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:07.551 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:35:07.551 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:07.551 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:07.551 [2024-11-20 13:55:14.605945] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:35:07.551 [2024-11-20 13:55:14.606014] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:35:07.551 [2024-11-20 13:55:14.606038] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:35:07.551 [2024-11-20 13:55:14.606051] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:35:07.551 [2024-11-20 13:55:14.608052] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:35:07.551 [2024-11-20 13:55:14.608087] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:35:07.810 BaseBdev1 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev2_malloc 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:07.810 BaseBdev2_malloc 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:07.810 [2024-11-20 13:55:14.644680] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:35:07.810 [2024-11-20 13:55:14.644741] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:35:07.810 [2024-11-20 13:55:14.644760] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:35:07.810 [2024-11-20 13:55:14.644773] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:35:07.810 [2024-11-20 13:55:14.646771] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:35:07.810 [2024-11-20 13:55:14.646806] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:35:07.810 BaseBdev2 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b spare_malloc 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:07.810 spare_malloc 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:07.810 spare_delay 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:07.810 [2024-11-20 13:55:14.704058] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:35:07.810 [2024-11-20 13:55:14.704119] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:35:07.810 [2024-11-20 13:55:14.704139] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:35:07.810 [2024-11-20 13:55:14.704151] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:35:07.810 [2024-11-20 13:55:14.706194] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:35:07.810 [2024-11-20 13:55:14.706228] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:35:07.810 spare 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:07.810 [2024-11-20 13:55:14.712105] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:35:07.810 [2024-11-20 13:55:14.714074] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:35:07.810 [2024-11-20 13:55:14.714242] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:35:07.810 [2024-11-20 13:55:14.714256] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:35:07.810 [2024-11-20 13:55:14.714334] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:35:07.810 [2024-11-20 13:55:14.714458] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:35:07.810 [2024-11-20 13:55:14.714469] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:35:07.810 [2024-11-20 13:55:14.714562] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:07.810 "name": "raid_bdev1", 00:35:07.810 "uuid": "554b0a5d-2255-4d67-8299-231e005d49b5", 00:35:07.810 "strip_size_kb": 0, 00:35:07.810 "state": "online", 00:35:07.810 "raid_level": "raid1", 00:35:07.810 "superblock": true, 00:35:07.810 "num_base_bdevs": 2, 00:35:07.810 "num_base_bdevs_discovered": 2, 00:35:07.810 "num_base_bdevs_operational": 2, 00:35:07.810 "base_bdevs_list": [ 00:35:07.810 { 00:35:07.810 "name": "BaseBdev1", 00:35:07.810 "uuid": "c53c599a-d23c-5c91-8334-88c8dbbf1ef8", 00:35:07.810 "is_configured": true, 00:35:07.810 "data_offset": 256, 00:35:07.810 "data_size": 7936 00:35:07.810 }, 00:35:07.810 { 00:35:07.810 "name": "BaseBdev2", 00:35:07.810 "uuid": "8d3cd44f-9051-537b-b56e-5bafca80b145", 00:35:07.810 "is_configured": true, 00:35:07.810 "data_offset": 256, 00:35:07.810 "data_size": 7936 00:35:07.810 } 00:35:07.810 ] 00:35:07.810 }' 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:07.810 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:08.069 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:35:08.069 13:55:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:35:08.069 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:08.069 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:08.069 [2024-11-20 13:55:15.008508] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:35:08.069 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:08.069 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:35:08.069 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:08.069 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:35:08.069 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:08.069 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:08.069 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:08.069 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:35:08.069 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:35:08.069 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:35:08.069 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:35:08.069 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:35:08.069 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:35:08.069 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:35:08.069 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # local bdev_list 00:35:08.069 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:35:08.069 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # local nbd_list 00:35:08.069 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@12 -- # local i 00:35:08.069 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:35:08.069 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:35:08.069 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:35:08.327 [2024-11-20 13:55:15.252317] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:35:08.327 /dev/nbd0 00:35:08.327 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:35:08.327 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:35:08.327 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:35:08.327 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # local i 00:35:08.327 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:35:08.327 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:35:08.327 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:35:08.327 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@877 -- # break 00:35:08.327 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:35:08.327 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:35:08.327 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:35:08.327 1+0 records in 00:35:08.327 1+0 records out 00:35:08.327 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00026965 s, 15.2 MB/s 00:35:08.327 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:35:08.327 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # size=4096 00:35:08.327 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:35:08.327 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:35:08.327 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@893 -- # return 0 00:35:08.327 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:35:08.327 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:35:08.327 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:35:08.327 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:35:08.327 13:55:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=4096 count=7936 oflag=direct 00:35:09.262 7936+0 records in 00:35:09.262 7936+0 records out 00:35:09.262 32505856 bytes (33 MB, 31 MiB) copied, 0.791445 s, 41.1 MB/s 00:35:09.262 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:35:09.262 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:35:09.262 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:35:09.262 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # local nbd_list 00:35:09.262 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@51 -- # local i 00:35:09.262 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:35:09.262 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:35:09.262 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:35:09.262 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:35:09.262 [2024-11-20 13:55:16.318147] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:35:09.262 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:35:09.262 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:35:09.262 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:35:09.262 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:35:09.520 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:35:09.520 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:35:09.520 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:35:09.520 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:09.520 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:09.520 [2024-11-20 13:55:16.330249] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:35:09.520 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:09.521 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:35:09.521 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:09.521 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:09.521 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:09.521 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:09.521 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:35:09.521 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:09.521 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:09.521 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:09.521 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:09.521 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:09.521 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:09.521 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:09.521 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:09.521 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:09.521 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:09.521 "name": "raid_bdev1", 00:35:09.521 "uuid": "554b0a5d-2255-4d67-8299-231e005d49b5", 00:35:09.521 "strip_size_kb": 0, 00:35:09.521 "state": "online", 00:35:09.521 "raid_level": "raid1", 00:35:09.521 "superblock": true, 00:35:09.521 "num_base_bdevs": 2, 00:35:09.521 "num_base_bdevs_discovered": 1, 00:35:09.521 "num_base_bdevs_operational": 1, 00:35:09.521 "base_bdevs_list": [ 00:35:09.521 { 00:35:09.521 "name": null, 00:35:09.521 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:09.521 "is_configured": false, 00:35:09.521 "data_offset": 0, 00:35:09.521 "data_size": 7936 00:35:09.521 }, 00:35:09.521 { 00:35:09.521 "name": "BaseBdev2", 00:35:09.521 "uuid": "8d3cd44f-9051-537b-b56e-5bafca80b145", 00:35:09.521 "is_configured": true, 00:35:09.521 "data_offset": 256, 00:35:09.521 "data_size": 7936 00:35:09.521 } 00:35:09.521 ] 00:35:09.521 }' 00:35:09.521 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:09.521 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:09.779 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:35:09.779 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:09.779 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:09.779 [2024-11-20 13:55:16.654355] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:35:09.779 [2024-11-20 13:55:16.664436] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00018d260 00:35:09.779 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:09.779 13:55:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@647 -- # sleep 1 00:35:09.779 [2024-11-20 13:55:16.666429] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:35:10.715 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:35:10.715 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:35:10.715 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:35:10.715 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:35:10.715 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:35:10.715 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:10.715 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:10.715 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:10.715 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:10.715 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:10.715 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:35:10.715 "name": "raid_bdev1", 00:35:10.715 "uuid": "554b0a5d-2255-4d67-8299-231e005d49b5", 00:35:10.715 "strip_size_kb": 0, 00:35:10.715 "state": "online", 00:35:10.715 "raid_level": "raid1", 00:35:10.715 "superblock": true, 00:35:10.715 "num_base_bdevs": 2, 00:35:10.715 "num_base_bdevs_discovered": 2, 00:35:10.715 "num_base_bdevs_operational": 2, 00:35:10.715 "process": { 00:35:10.715 "type": "rebuild", 00:35:10.715 "target": "spare", 00:35:10.715 "progress": { 00:35:10.715 "blocks": 2560, 00:35:10.715 "percent": 32 00:35:10.715 } 00:35:10.715 }, 00:35:10.715 "base_bdevs_list": [ 00:35:10.715 { 00:35:10.715 "name": "spare", 00:35:10.715 "uuid": "e0a797ae-d999-58ac-8b3c-f49f4dbbb5d6", 00:35:10.715 "is_configured": true, 00:35:10.715 "data_offset": 256, 00:35:10.715 "data_size": 7936 00:35:10.715 }, 00:35:10.715 { 00:35:10.715 "name": "BaseBdev2", 00:35:10.715 "uuid": "8d3cd44f-9051-537b-b56e-5bafca80b145", 00:35:10.715 "is_configured": true, 00:35:10.715 "data_offset": 256, 00:35:10.715 "data_size": 7936 00:35:10.715 } 00:35:10.715 ] 00:35:10.715 }' 00:35:10.715 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:35:10.715 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:35:10.715 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:35:10.715 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:35:10.715 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:35:10.715 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:10.715 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:10.974 [2024-11-20 13:55:17.776517] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:35:10.974 [2024-11-20 13:55:17.873613] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:35:10.974 [2024-11-20 13:55:17.873689] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:35:10.974 [2024-11-20 13:55:17.873703] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:35:10.974 [2024-11-20 13:55:17.873712] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:35:10.974 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:10.974 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:35:10.974 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:10.974 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:10.974 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:10.974 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:10.974 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:35:10.974 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:10.974 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:10.974 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:10.974 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:10.974 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:10.974 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:10.974 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:10.974 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:10.974 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:10.974 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:10.974 "name": "raid_bdev1", 00:35:10.974 "uuid": "554b0a5d-2255-4d67-8299-231e005d49b5", 00:35:10.974 "strip_size_kb": 0, 00:35:10.974 "state": "online", 00:35:10.974 "raid_level": "raid1", 00:35:10.974 "superblock": true, 00:35:10.974 "num_base_bdevs": 2, 00:35:10.974 "num_base_bdevs_discovered": 1, 00:35:10.974 "num_base_bdevs_operational": 1, 00:35:10.974 "base_bdevs_list": [ 00:35:10.974 { 00:35:10.974 "name": null, 00:35:10.974 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:10.974 "is_configured": false, 00:35:10.974 "data_offset": 0, 00:35:10.974 "data_size": 7936 00:35:10.974 }, 00:35:10.974 { 00:35:10.974 "name": "BaseBdev2", 00:35:10.974 "uuid": "8d3cd44f-9051-537b-b56e-5bafca80b145", 00:35:10.974 "is_configured": true, 00:35:10.974 "data_offset": 256, 00:35:10.974 "data_size": 7936 00:35:10.974 } 00:35:10.974 ] 00:35:10.974 }' 00:35:10.974 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:10.974 13:55:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:11.232 13:55:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:35:11.232 13:55:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:35:11.232 13:55:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:35:11.232 13:55:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:35:11.232 13:55:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:35:11.232 13:55:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:11.232 13:55:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:11.232 13:55:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:11.232 13:55:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:11.232 13:55:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:11.232 13:55:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:35:11.232 "name": "raid_bdev1", 00:35:11.232 "uuid": "554b0a5d-2255-4d67-8299-231e005d49b5", 00:35:11.232 "strip_size_kb": 0, 00:35:11.232 "state": "online", 00:35:11.232 "raid_level": "raid1", 00:35:11.232 "superblock": true, 00:35:11.232 "num_base_bdevs": 2, 00:35:11.232 "num_base_bdevs_discovered": 1, 00:35:11.232 "num_base_bdevs_operational": 1, 00:35:11.232 "base_bdevs_list": [ 00:35:11.232 { 00:35:11.232 "name": null, 00:35:11.232 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:11.232 "is_configured": false, 00:35:11.232 "data_offset": 0, 00:35:11.232 "data_size": 7936 00:35:11.232 }, 00:35:11.232 { 00:35:11.232 "name": "BaseBdev2", 00:35:11.232 "uuid": "8d3cd44f-9051-537b-b56e-5bafca80b145", 00:35:11.232 "is_configured": true, 00:35:11.232 "data_offset": 256, 00:35:11.232 "data_size": 7936 00:35:11.232 } 00:35:11.232 ] 00:35:11.232 }' 00:35:11.232 13:55:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:35:11.232 13:55:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:35:11.232 13:55:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:35:11.490 13:55:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:35:11.490 13:55:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:35:11.490 13:55:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:11.490 13:55:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:11.490 [2024-11-20 13:55:18.318351] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:35:11.490 [2024-11-20 13:55:18.325688] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00018d330 00:35:11.490 13:55:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:11.490 13:55:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@663 -- # sleep 1 00:35:11.491 [2024-11-20 13:55:18.327345] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:35:12.424 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:35:12.424 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:35:12.424 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:35:12.424 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:35:12.424 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:35:12.424 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:12.424 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:12.424 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:12.424 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:12.424 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:12.424 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:35:12.424 "name": "raid_bdev1", 00:35:12.424 "uuid": "554b0a5d-2255-4d67-8299-231e005d49b5", 00:35:12.424 "strip_size_kb": 0, 00:35:12.424 "state": "online", 00:35:12.424 "raid_level": "raid1", 00:35:12.424 "superblock": true, 00:35:12.424 "num_base_bdevs": 2, 00:35:12.424 "num_base_bdevs_discovered": 2, 00:35:12.424 "num_base_bdevs_operational": 2, 00:35:12.424 "process": { 00:35:12.424 "type": "rebuild", 00:35:12.424 "target": "spare", 00:35:12.424 "progress": { 00:35:12.424 "blocks": 2560, 00:35:12.424 "percent": 32 00:35:12.424 } 00:35:12.424 }, 00:35:12.424 "base_bdevs_list": [ 00:35:12.424 { 00:35:12.424 "name": "spare", 00:35:12.424 "uuid": "e0a797ae-d999-58ac-8b3c-f49f4dbbb5d6", 00:35:12.424 "is_configured": true, 00:35:12.424 "data_offset": 256, 00:35:12.424 "data_size": 7936 00:35:12.424 }, 00:35:12.424 { 00:35:12.424 "name": "BaseBdev2", 00:35:12.424 "uuid": "8d3cd44f-9051-537b-b56e-5bafca80b145", 00:35:12.424 "is_configured": true, 00:35:12.424 "data_offset": 256, 00:35:12.424 "data_size": 7936 00:35:12.424 } 00:35:12.424 ] 00:35:12.424 }' 00:35:12.424 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:35:12.424 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:35:12.424 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:35:12.424 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:35:12.424 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:35:12.424 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:35:12.424 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:35:12.424 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:35:12.424 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:35:12.424 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:35:12.424 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@706 -- # local timeout=566 00:35:12.424 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:35:12.424 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:35:12.424 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:35:12.424 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:35:12.425 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:35:12.425 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:35:12.425 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:12.425 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:12.425 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:12.425 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:12.425 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:12.425 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:35:12.425 "name": "raid_bdev1", 00:35:12.425 "uuid": "554b0a5d-2255-4d67-8299-231e005d49b5", 00:35:12.425 "strip_size_kb": 0, 00:35:12.425 "state": "online", 00:35:12.425 "raid_level": "raid1", 00:35:12.425 "superblock": true, 00:35:12.425 "num_base_bdevs": 2, 00:35:12.425 "num_base_bdevs_discovered": 2, 00:35:12.425 "num_base_bdevs_operational": 2, 00:35:12.425 "process": { 00:35:12.425 "type": "rebuild", 00:35:12.425 "target": "spare", 00:35:12.425 "progress": { 00:35:12.425 "blocks": 2816, 00:35:12.425 "percent": 35 00:35:12.425 } 00:35:12.425 }, 00:35:12.425 "base_bdevs_list": [ 00:35:12.425 { 00:35:12.425 "name": "spare", 00:35:12.425 "uuid": "e0a797ae-d999-58ac-8b3c-f49f4dbbb5d6", 00:35:12.425 "is_configured": true, 00:35:12.425 "data_offset": 256, 00:35:12.425 "data_size": 7936 00:35:12.425 }, 00:35:12.425 { 00:35:12.425 "name": "BaseBdev2", 00:35:12.425 "uuid": "8d3cd44f-9051-537b-b56e-5bafca80b145", 00:35:12.425 "is_configured": true, 00:35:12.425 "data_offset": 256, 00:35:12.425 "data_size": 7936 00:35:12.425 } 00:35:12.425 ] 00:35:12.425 }' 00:35:12.425 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:35:12.683 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:35:12.683 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:35:12.683 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:35:12.683 13:55:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@711 -- # sleep 1 00:35:13.618 13:55:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:35:13.618 13:55:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:35:13.618 13:55:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:35:13.618 13:55:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:35:13.618 13:55:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:35:13.618 13:55:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:35:13.618 13:55:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:13.618 13:55:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:13.618 13:55:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:13.618 13:55:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:13.618 13:55:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:13.618 13:55:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:35:13.618 "name": "raid_bdev1", 00:35:13.618 "uuid": "554b0a5d-2255-4d67-8299-231e005d49b5", 00:35:13.618 "strip_size_kb": 0, 00:35:13.618 "state": "online", 00:35:13.618 "raid_level": "raid1", 00:35:13.618 "superblock": true, 00:35:13.618 "num_base_bdevs": 2, 00:35:13.618 "num_base_bdevs_discovered": 2, 00:35:13.618 "num_base_bdevs_operational": 2, 00:35:13.618 "process": { 00:35:13.618 "type": "rebuild", 00:35:13.618 "target": "spare", 00:35:13.618 "progress": { 00:35:13.618 "blocks": 5632, 00:35:13.618 "percent": 70 00:35:13.618 } 00:35:13.618 }, 00:35:13.618 "base_bdevs_list": [ 00:35:13.618 { 00:35:13.618 "name": "spare", 00:35:13.618 "uuid": "e0a797ae-d999-58ac-8b3c-f49f4dbbb5d6", 00:35:13.618 "is_configured": true, 00:35:13.618 "data_offset": 256, 00:35:13.618 "data_size": 7936 00:35:13.618 }, 00:35:13.618 { 00:35:13.618 "name": "BaseBdev2", 00:35:13.618 "uuid": "8d3cd44f-9051-537b-b56e-5bafca80b145", 00:35:13.618 "is_configured": true, 00:35:13.618 "data_offset": 256, 00:35:13.618 "data_size": 7936 00:35:13.618 } 00:35:13.618 ] 00:35:13.618 }' 00:35:13.618 13:55:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:35:13.618 13:55:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:35:13.618 13:55:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:35:13.618 13:55:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:35:13.618 13:55:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@711 -- # sleep 1 00:35:14.552 [2024-11-20 13:55:21.444397] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:35:14.552 [2024-11-20 13:55:21.444470] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:35:14.552 [2024-11-20 13:55:21.444567] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:35:14.811 "name": "raid_bdev1", 00:35:14.811 "uuid": "554b0a5d-2255-4d67-8299-231e005d49b5", 00:35:14.811 "strip_size_kb": 0, 00:35:14.811 "state": "online", 00:35:14.811 "raid_level": "raid1", 00:35:14.811 "superblock": true, 00:35:14.811 "num_base_bdevs": 2, 00:35:14.811 "num_base_bdevs_discovered": 2, 00:35:14.811 "num_base_bdevs_operational": 2, 00:35:14.811 "base_bdevs_list": [ 00:35:14.811 { 00:35:14.811 "name": "spare", 00:35:14.811 "uuid": "e0a797ae-d999-58ac-8b3c-f49f4dbbb5d6", 00:35:14.811 "is_configured": true, 00:35:14.811 "data_offset": 256, 00:35:14.811 "data_size": 7936 00:35:14.811 }, 00:35:14.811 { 00:35:14.811 "name": "BaseBdev2", 00:35:14.811 "uuid": "8d3cd44f-9051-537b-b56e-5bafca80b145", 00:35:14.811 "is_configured": true, 00:35:14.811 "data_offset": 256, 00:35:14.811 "data_size": 7936 00:35:14.811 } 00:35:14.811 ] 00:35:14.811 }' 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@709 -- # break 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:35:14.811 "name": "raid_bdev1", 00:35:14.811 "uuid": "554b0a5d-2255-4d67-8299-231e005d49b5", 00:35:14.811 "strip_size_kb": 0, 00:35:14.811 "state": "online", 00:35:14.811 "raid_level": "raid1", 00:35:14.811 "superblock": true, 00:35:14.811 "num_base_bdevs": 2, 00:35:14.811 "num_base_bdevs_discovered": 2, 00:35:14.811 "num_base_bdevs_operational": 2, 00:35:14.811 "base_bdevs_list": [ 00:35:14.811 { 00:35:14.811 "name": "spare", 00:35:14.811 "uuid": "e0a797ae-d999-58ac-8b3c-f49f4dbbb5d6", 00:35:14.811 "is_configured": true, 00:35:14.811 "data_offset": 256, 00:35:14.811 "data_size": 7936 00:35:14.811 }, 00:35:14.811 { 00:35:14.811 "name": "BaseBdev2", 00:35:14.811 "uuid": "8d3cd44f-9051-537b-b56e-5bafca80b145", 00:35:14.811 "is_configured": true, 00:35:14.811 "data_offset": 256, 00:35:14.811 "data_size": 7936 00:35:14.811 } 00:35:14.811 ] 00:35:14.811 }' 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:14.811 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:14.811 "name": "raid_bdev1", 00:35:14.812 "uuid": "554b0a5d-2255-4d67-8299-231e005d49b5", 00:35:14.812 "strip_size_kb": 0, 00:35:14.812 "state": "online", 00:35:14.812 "raid_level": "raid1", 00:35:14.812 "superblock": true, 00:35:14.812 "num_base_bdevs": 2, 00:35:14.812 "num_base_bdevs_discovered": 2, 00:35:14.812 "num_base_bdevs_operational": 2, 00:35:14.812 "base_bdevs_list": [ 00:35:14.812 { 00:35:14.812 "name": "spare", 00:35:14.812 "uuid": "e0a797ae-d999-58ac-8b3c-f49f4dbbb5d6", 00:35:14.812 "is_configured": true, 00:35:14.812 "data_offset": 256, 00:35:14.812 "data_size": 7936 00:35:14.812 }, 00:35:14.812 { 00:35:14.812 "name": "BaseBdev2", 00:35:14.812 "uuid": "8d3cd44f-9051-537b-b56e-5bafca80b145", 00:35:14.812 "is_configured": true, 00:35:14.812 "data_offset": 256, 00:35:14.812 "data_size": 7936 00:35:14.812 } 00:35:14.812 ] 00:35:14.812 }' 00:35:14.812 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:14.812 13:55:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:15.069 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:35:15.069 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:15.069 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:15.069 [2024-11-20 13:55:22.117297] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:35:15.069 [2024-11-20 13:55:22.117329] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:35:15.069 [2024-11-20 13:55:22.117407] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:35:15.069 [2024-11-20 13:55:22.117480] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:35:15.069 [2024-11-20 13:55:22.117490] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:35:15.069 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:15.069 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:15.069 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:15.069 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:15.070 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # jq length 00:35:15.328 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:15.328 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:35:15.328 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:35:15.328 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:35:15.328 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:35:15.328 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:35:15.328 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:35:15.328 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # local bdev_list 00:35:15.328 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:35:15.328 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # local nbd_list 00:35:15.328 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@12 -- # local i 00:35:15.328 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:35:15.328 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:35:15.328 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:35:15.328 /dev/nbd0 00:35:15.328 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:35:15.328 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:35:15.328 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:35:15.328 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # local i 00:35:15.328 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:35:15.328 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:35:15.328 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@877 -- # break 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:35:15.587 1+0 records in 00:35:15.587 1+0 records out 00:35:15.587 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00017174 s, 23.9 MB/s 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # size=4096 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@893 -- # return 0 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:35:15.587 /dev/nbd1 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # local i 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@877 -- # break 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:35:15.587 1+0 records in 00:35:15.587 1+0 records out 00:35:15.587 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000326349 s, 12.6 MB/s 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # size=4096 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@893 -- # return 0 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:35:15.587 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:35:15.845 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:35:15.845 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:35:15.845 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:35:15.845 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # local nbd_list 00:35:15.845 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@51 -- # local i 00:35:15.845 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:35:15.845 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:35:16.104 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:35:16.104 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:35:16.104 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:35:16.104 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:35:16.104 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:35:16.104 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:35:16.104 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:35:16.104 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:35:16.104 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:35:16.104 13:55:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:16.362 [2024-11-20 13:55:23.244769] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:35:16.362 [2024-11-20 13:55:23.244820] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:35:16.362 [2024-11-20 13:55:23.244843] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:35:16.362 [2024-11-20 13:55:23.244852] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:35:16.362 [2024-11-20 13:55:23.246674] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:35:16.362 [2024-11-20 13:55:23.246707] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:35:16.362 [2024-11-20 13:55:23.246765] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:35:16.362 [2024-11-20 13:55:23.246807] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:35:16.362 [2024-11-20 13:55:23.246921] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:35:16.362 spare 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:16.362 [2024-11-20 13:55:23.347007] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:35:16.362 [2024-11-20 13:55:23.347033] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:35:16.362 [2024-11-20 13:55:23.347124] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1b50 00:35:16.362 [2024-11-20 13:55:23.347241] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:35:16.362 [2024-11-20 13:55:23.347257] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:35:16.362 [2024-11-20 13:55:23.347353] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:16.362 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:16.362 "name": "raid_bdev1", 00:35:16.362 "uuid": "554b0a5d-2255-4d67-8299-231e005d49b5", 00:35:16.362 "strip_size_kb": 0, 00:35:16.362 "state": "online", 00:35:16.362 "raid_level": "raid1", 00:35:16.362 "superblock": true, 00:35:16.362 "num_base_bdevs": 2, 00:35:16.362 "num_base_bdevs_discovered": 2, 00:35:16.362 "num_base_bdevs_operational": 2, 00:35:16.362 "base_bdevs_list": [ 00:35:16.362 { 00:35:16.362 "name": "spare", 00:35:16.362 "uuid": "e0a797ae-d999-58ac-8b3c-f49f4dbbb5d6", 00:35:16.362 "is_configured": true, 00:35:16.362 "data_offset": 256, 00:35:16.362 "data_size": 7936 00:35:16.362 }, 00:35:16.362 { 00:35:16.362 "name": "BaseBdev2", 00:35:16.362 "uuid": "8d3cd44f-9051-537b-b56e-5bafca80b145", 00:35:16.363 "is_configured": true, 00:35:16.363 "data_offset": 256, 00:35:16.363 "data_size": 7936 00:35:16.363 } 00:35:16.363 ] 00:35:16.363 }' 00:35:16.363 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:16.363 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:16.620 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:35:16.620 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:35:16.620 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:35:16.620 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:35:16.620 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:35:16.620 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:16.620 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:16.620 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:16.620 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:16.878 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:16.878 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:35:16.878 "name": "raid_bdev1", 00:35:16.878 "uuid": "554b0a5d-2255-4d67-8299-231e005d49b5", 00:35:16.878 "strip_size_kb": 0, 00:35:16.878 "state": "online", 00:35:16.878 "raid_level": "raid1", 00:35:16.878 "superblock": true, 00:35:16.878 "num_base_bdevs": 2, 00:35:16.878 "num_base_bdevs_discovered": 2, 00:35:16.878 "num_base_bdevs_operational": 2, 00:35:16.878 "base_bdevs_list": [ 00:35:16.878 { 00:35:16.878 "name": "spare", 00:35:16.878 "uuid": "e0a797ae-d999-58ac-8b3c-f49f4dbbb5d6", 00:35:16.878 "is_configured": true, 00:35:16.878 "data_offset": 256, 00:35:16.878 "data_size": 7936 00:35:16.878 }, 00:35:16.878 { 00:35:16.878 "name": "BaseBdev2", 00:35:16.878 "uuid": "8d3cd44f-9051-537b-b56e-5bafca80b145", 00:35:16.879 "is_configured": true, 00:35:16.879 "data_offset": 256, 00:35:16.879 "data_size": 7936 00:35:16.879 } 00:35:16.879 ] 00:35:16.879 }' 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:16.879 [2024-11-20 13:55:23.804905] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:16.879 "name": "raid_bdev1", 00:35:16.879 "uuid": "554b0a5d-2255-4d67-8299-231e005d49b5", 00:35:16.879 "strip_size_kb": 0, 00:35:16.879 "state": "online", 00:35:16.879 "raid_level": "raid1", 00:35:16.879 "superblock": true, 00:35:16.879 "num_base_bdevs": 2, 00:35:16.879 "num_base_bdevs_discovered": 1, 00:35:16.879 "num_base_bdevs_operational": 1, 00:35:16.879 "base_bdevs_list": [ 00:35:16.879 { 00:35:16.879 "name": null, 00:35:16.879 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:16.879 "is_configured": false, 00:35:16.879 "data_offset": 0, 00:35:16.879 "data_size": 7936 00:35:16.879 }, 00:35:16.879 { 00:35:16.879 "name": "BaseBdev2", 00:35:16.879 "uuid": "8d3cd44f-9051-537b-b56e-5bafca80b145", 00:35:16.879 "is_configured": true, 00:35:16.879 "data_offset": 256, 00:35:16.879 "data_size": 7936 00:35:16.879 } 00:35:16.879 ] 00:35:16.879 }' 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:16.879 13:55:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:17.137 13:55:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:35:17.137 13:55:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:17.137 13:55:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:17.137 [2024-11-20 13:55:24.116994] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:35:17.137 [2024-11-20 13:55:24.117175] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:35:17.137 [2024-11-20 13:55:24.117194] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:35:17.137 [2024-11-20 13:55:24.117228] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:35:17.137 [2024-11-20 13:55:24.124064] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1c20 00:35:17.137 13:55:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:17.137 13:55:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@757 -- # sleep 1 00:35:17.137 [2024-11-20 13:55:24.125721] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:35:18.072 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:35:18.072 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:35:18.072 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:35:18.072 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:35:18.072 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:35:18.330 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:18.330 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:18.330 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:18.330 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:18.330 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:18.330 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:35:18.330 "name": "raid_bdev1", 00:35:18.330 "uuid": "554b0a5d-2255-4d67-8299-231e005d49b5", 00:35:18.330 "strip_size_kb": 0, 00:35:18.330 "state": "online", 00:35:18.330 "raid_level": "raid1", 00:35:18.330 "superblock": true, 00:35:18.330 "num_base_bdevs": 2, 00:35:18.330 "num_base_bdevs_discovered": 2, 00:35:18.331 "num_base_bdevs_operational": 2, 00:35:18.331 "process": { 00:35:18.331 "type": "rebuild", 00:35:18.331 "target": "spare", 00:35:18.331 "progress": { 00:35:18.331 "blocks": 2560, 00:35:18.331 "percent": 32 00:35:18.331 } 00:35:18.331 }, 00:35:18.331 "base_bdevs_list": [ 00:35:18.331 { 00:35:18.331 "name": "spare", 00:35:18.331 "uuid": "e0a797ae-d999-58ac-8b3c-f49f4dbbb5d6", 00:35:18.331 "is_configured": true, 00:35:18.331 "data_offset": 256, 00:35:18.331 "data_size": 7936 00:35:18.331 }, 00:35:18.331 { 00:35:18.331 "name": "BaseBdev2", 00:35:18.331 "uuid": "8d3cd44f-9051-537b-b56e-5bafca80b145", 00:35:18.331 "is_configured": true, 00:35:18.331 "data_offset": 256, 00:35:18.331 "data_size": 7936 00:35:18.331 } 00:35:18.331 ] 00:35:18.331 }' 00:35:18.331 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:35:18.331 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:35:18.331 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:35:18.331 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:35:18.331 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:35:18.331 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:18.331 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:18.331 [2024-11-20 13:55:25.240291] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:35:18.331 [2024-11-20 13:55:25.332215] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:35:18.331 [2024-11-20 13:55:25.332274] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:35:18.331 [2024-11-20 13:55:25.332286] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:35:18.331 [2024-11-20 13:55:25.332300] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:35:18.331 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:18.331 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:35:18.331 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:18.331 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:18.331 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:18.331 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:18.331 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:35:18.331 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:18.331 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:18.331 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:18.331 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:18.331 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:18.331 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:18.331 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:18.331 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:18.331 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:18.331 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:18.331 "name": "raid_bdev1", 00:35:18.331 "uuid": "554b0a5d-2255-4d67-8299-231e005d49b5", 00:35:18.331 "strip_size_kb": 0, 00:35:18.331 "state": "online", 00:35:18.331 "raid_level": "raid1", 00:35:18.331 "superblock": true, 00:35:18.331 "num_base_bdevs": 2, 00:35:18.331 "num_base_bdevs_discovered": 1, 00:35:18.331 "num_base_bdevs_operational": 1, 00:35:18.331 "base_bdevs_list": [ 00:35:18.331 { 00:35:18.331 "name": null, 00:35:18.331 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:18.331 "is_configured": false, 00:35:18.331 "data_offset": 0, 00:35:18.331 "data_size": 7936 00:35:18.331 }, 00:35:18.331 { 00:35:18.331 "name": "BaseBdev2", 00:35:18.331 "uuid": "8d3cd44f-9051-537b-b56e-5bafca80b145", 00:35:18.331 "is_configured": true, 00:35:18.331 "data_offset": 256, 00:35:18.331 "data_size": 7936 00:35:18.331 } 00:35:18.331 ] 00:35:18.331 }' 00:35:18.331 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:18.331 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:18.897 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:35:18.897 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:18.897 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:18.897 [2024-11-20 13:55:25.660739] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:35:18.897 [2024-11-20 13:55:25.660799] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:35:18.897 [2024-11-20 13:55:25.660822] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:35:18.897 [2024-11-20 13:55:25.660833] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:35:18.898 [2024-11-20 13:55:25.661078] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:35:18.898 [2024-11-20 13:55:25.661097] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:35:18.898 [2024-11-20 13:55:25.661148] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:35:18.898 [2024-11-20 13:55:25.661160] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:35:18.898 [2024-11-20 13:55:25.661169] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:35:18.898 [2024-11-20 13:55:25.661185] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:35:18.898 [2024-11-20 13:55:25.668170] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1cf0 00:35:18.898 spare 00:35:18.898 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:18.898 13:55:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@764 -- # sleep 1 00:35:18.898 [2024-11-20 13:55:25.669805] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:35:19.832 "name": "raid_bdev1", 00:35:19.832 "uuid": "554b0a5d-2255-4d67-8299-231e005d49b5", 00:35:19.832 "strip_size_kb": 0, 00:35:19.832 "state": "online", 00:35:19.832 "raid_level": "raid1", 00:35:19.832 "superblock": true, 00:35:19.832 "num_base_bdevs": 2, 00:35:19.832 "num_base_bdevs_discovered": 2, 00:35:19.832 "num_base_bdevs_operational": 2, 00:35:19.832 "process": { 00:35:19.832 "type": "rebuild", 00:35:19.832 "target": "spare", 00:35:19.832 "progress": { 00:35:19.832 "blocks": 2560, 00:35:19.832 "percent": 32 00:35:19.832 } 00:35:19.832 }, 00:35:19.832 "base_bdevs_list": [ 00:35:19.832 { 00:35:19.832 "name": "spare", 00:35:19.832 "uuid": "e0a797ae-d999-58ac-8b3c-f49f4dbbb5d6", 00:35:19.832 "is_configured": true, 00:35:19.832 "data_offset": 256, 00:35:19.832 "data_size": 7936 00:35:19.832 }, 00:35:19.832 { 00:35:19.832 "name": "BaseBdev2", 00:35:19.832 "uuid": "8d3cd44f-9051-537b-b56e-5bafca80b145", 00:35:19.832 "is_configured": true, 00:35:19.832 "data_offset": 256, 00:35:19.832 "data_size": 7936 00:35:19.832 } 00:35:19.832 ] 00:35:19.832 }' 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:19.832 [2024-11-20 13:55:26.768530] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:35:19.832 [2024-11-20 13:55:26.776143] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:35:19.832 [2024-11-20 13:55:26.776195] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:35:19.832 [2024-11-20 13:55:26.776209] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:35:19.832 [2024-11-20 13:55:26.776216] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:19.832 "name": "raid_bdev1", 00:35:19.832 "uuid": "554b0a5d-2255-4d67-8299-231e005d49b5", 00:35:19.832 "strip_size_kb": 0, 00:35:19.832 "state": "online", 00:35:19.832 "raid_level": "raid1", 00:35:19.832 "superblock": true, 00:35:19.832 "num_base_bdevs": 2, 00:35:19.832 "num_base_bdevs_discovered": 1, 00:35:19.832 "num_base_bdevs_operational": 1, 00:35:19.832 "base_bdevs_list": [ 00:35:19.832 { 00:35:19.832 "name": null, 00:35:19.832 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:19.832 "is_configured": false, 00:35:19.832 "data_offset": 0, 00:35:19.832 "data_size": 7936 00:35:19.832 }, 00:35:19.832 { 00:35:19.832 "name": "BaseBdev2", 00:35:19.832 "uuid": "8d3cd44f-9051-537b-b56e-5bafca80b145", 00:35:19.832 "is_configured": true, 00:35:19.832 "data_offset": 256, 00:35:19.832 "data_size": 7936 00:35:19.832 } 00:35:19.832 ] 00:35:19.832 }' 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:19.832 13:55:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:20.091 13:55:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:35:20.091 13:55:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:35:20.091 13:55:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:35:20.091 13:55:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:35:20.091 13:55:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:35:20.091 13:55:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:20.091 13:55:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:20.091 13:55:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:20.091 13:55:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:20.091 13:55:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:20.350 13:55:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:35:20.350 "name": "raid_bdev1", 00:35:20.350 "uuid": "554b0a5d-2255-4d67-8299-231e005d49b5", 00:35:20.350 "strip_size_kb": 0, 00:35:20.350 "state": "online", 00:35:20.350 "raid_level": "raid1", 00:35:20.350 "superblock": true, 00:35:20.350 "num_base_bdevs": 2, 00:35:20.350 "num_base_bdevs_discovered": 1, 00:35:20.350 "num_base_bdevs_operational": 1, 00:35:20.350 "base_bdevs_list": [ 00:35:20.350 { 00:35:20.350 "name": null, 00:35:20.350 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:20.350 "is_configured": false, 00:35:20.350 "data_offset": 0, 00:35:20.350 "data_size": 7936 00:35:20.350 }, 00:35:20.350 { 00:35:20.350 "name": "BaseBdev2", 00:35:20.350 "uuid": "8d3cd44f-9051-537b-b56e-5bafca80b145", 00:35:20.350 "is_configured": true, 00:35:20.350 "data_offset": 256, 00:35:20.350 "data_size": 7936 00:35:20.350 } 00:35:20.350 ] 00:35:20.350 }' 00:35:20.350 13:55:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:35:20.350 13:55:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:35:20.350 13:55:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:35:20.350 13:55:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:35:20.350 13:55:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:35:20.350 13:55:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:20.350 13:55:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:20.350 13:55:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:20.350 13:55:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:35:20.350 13:55:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:20.350 13:55:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:20.350 [2024-11-20 13:55:27.221048] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:35:20.350 [2024-11-20 13:55:27.221098] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:35:20.350 [2024-11-20 13:55:27.221117] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:35:20.350 [2024-11-20 13:55:27.221126] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:35:20.350 [2024-11-20 13:55:27.221333] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:35:20.350 [2024-11-20 13:55:27.221348] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:35:20.350 [2024-11-20 13:55:27.221392] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:35:20.350 [2024-11-20 13:55:27.221403] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:35:20.350 [2024-11-20 13:55:27.221411] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:35:20.350 [2024-11-20 13:55:27.221419] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:35:20.350 BaseBdev1 00:35:20.350 13:55:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:20.350 13:55:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@775 -- # sleep 1 00:35:21.289 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:35:21.289 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:21.289 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:21.289 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:21.289 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:21.290 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:35:21.290 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:21.290 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:21.290 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:21.290 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:21.290 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:21.290 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:21.290 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:21.290 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:21.290 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:21.290 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:21.290 "name": "raid_bdev1", 00:35:21.290 "uuid": "554b0a5d-2255-4d67-8299-231e005d49b5", 00:35:21.290 "strip_size_kb": 0, 00:35:21.290 "state": "online", 00:35:21.290 "raid_level": "raid1", 00:35:21.290 "superblock": true, 00:35:21.290 "num_base_bdevs": 2, 00:35:21.290 "num_base_bdevs_discovered": 1, 00:35:21.290 "num_base_bdevs_operational": 1, 00:35:21.290 "base_bdevs_list": [ 00:35:21.290 { 00:35:21.290 "name": null, 00:35:21.290 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:21.290 "is_configured": false, 00:35:21.290 "data_offset": 0, 00:35:21.290 "data_size": 7936 00:35:21.290 }, 00:35:21.290 { 00:35:21.290 "name": "BaseBdev2", 00:35:21.290 "uuid": "8d3cd44f-9051-537b-b56e-5bafca80b145", 00:35:21.290 "is_configured": true, 00:35:21.290 "data_offset": 256, 00:35:21.290 "data_size": 7936 00:35:21.290 } 00:35:21.290 ] 00:35:21.290 }' 00:35:21.290 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:21.290 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:21.547 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:35:21.547 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:35:21.547 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:35:21.547 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:35:21.547 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:35:21.547 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:21.547 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:21.547 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:21.547 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:21.547 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:21.547 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:35:21.547 "name": "raid_bdev1", 00:35:21.547 "uuid": "554b0a5d-2255-4d67-8299-231e005d49b5", 00:35:21.547 "strip_size_kb": 0, 00:35:21.547 "state": "online", 00:35:21.547 "raid_level": "raid1", 00:35:21.547 "superblock": true, 00:35:21.547 "num_base_bdevs": 2, 00:35:21.547 "num_base_bdevs_discovered": 1, 00:35:21.547 "num_base_bdevs_operational": 1, 00:35:21.547 "base_bdevs_list": [ 00:35:21.547 { 00:35:21.547 "name": null, 00:35:21.547 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:21.547 "is_configured": false, 00:35:21.547 "data_offset": 0, 00:35:21.547 "data_size": 7936 00:35:21.547 }, 00:35:21.547 { 00:35:21.547 "name": "BaseBdev2", 00:35:21.547 "uuid": "8d3cd44f-9051-537b-b56e-5bafca80b145", 00:35:21.547 "is_configured": true, 00:35:21.547 "data_offset": 256, 00:35:21.547 "data_size": 7936 00:35:21.547 } 00:35:21.547 ] 00:35:21.547 }' 00:35:21.547 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:35:21.804 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:35:21.804 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:35:21.804 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:35:21.804 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:35:21.804 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@652 -- # local es=0 00:35:21.804 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:35:21.804 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:35:21.804 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:35:21.804 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:35:21.804 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:35:21.804 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:35:21.804 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:21.804 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:21.804 [2024-11-20 13:55:28.645346] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:35:21.804 [2024-11-20 13:55:28.645500] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:35:21.804 [2024-11-20 13:55:28.645520] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:35:21.804 request: 00:35:21.804 { 00:35:21.804 "base_bdev": "BaseBdev1", 00:35:21.804 "raid_bdev": "raid_bdev1", 00:35:21.804 "method": "bdev_raid_add_base_bdev", 00:35:21.804 "req_id": 1 00:35:21.804 } 00:35:21.804 Got JSON-RPC error response 00:35:21.804 response: 00:35:21.804 { 00:35:21.804 "code": -22, 00:35:21.804 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:35:21.804 } 00:35:21.804 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:35:21.804 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@655 -- # es=1 00:35:21.804 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:35:21.804 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:35:21.804 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:35:21.804 13:55:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@779 -- # sleep 1 00:35:22.760 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:35:22.760 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:22.760 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:22.760 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:22.760 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:22.760 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:35:22.760 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:22.760 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:22.760 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:22.760 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:22.760 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:22.760 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:22.760 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:22.760 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:22.760 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:22.760 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:22.760 "name": "raid_bdev1", 00:35:22.760 "uuid": "554b0a5d-2255-4d67-8299-231e005d49b5", 00:35:22.760 "strip_size_kb": 0, 00:35:22.760 "state": "online", 00:35:22.760 "raid_level": "raid1", 00:35:22.760 "superblock": true, 00:35:22.760 "num_base_bdevs": 2, 00:35:22.760 "num_base_bdevs_discovered": 1, 00:35:22.760 "num_base_bdevs_operational": 1, 00:35:22.760 "base_bdevs_list": [ 00:35:22.760 { 00:35:22.760 "name": null, 00:35:22.760 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:22.760 "is_configured": false, 00:35:22.760 "data_offset": 0, 00:35:22.760 "data_size": 7936 00:35:22.760 }, 00:35:22.760 { 00:35:22.760 "name": "BaseBdev2", 00:35:22.760 "uuid": "8d3cd44f-9051-537b-b56e-5bafca80b145", 00:35:22.760 "is_configured": true, 00:35:22.760 "data_offset": 256, 00:35:22.760 "data_size": 7936 00:35:22.760 } 00:35:22.760 ] 00:35:22.760 }' 00:35:22.760 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:22.760 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:23.019 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:35:23.019 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:35:23.019 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:35:23.019 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:35:23.019 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:35:23.019 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:23.019 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:23.019 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:23.019 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:23.019 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:23.019 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:35:23.019 "name": "raid_bdev1", 00:35:23.019 "uuid": "554b0a5d-2255-4d67-8299-231e005d49b5", 00:35:23.019 "strip_size_kb": 0, 00:35:23.019 "state": "online", 00:35:23.019 "raid_level": "raid1", 00:35:23.019 "superblock": true, 00:35:23.019 "num_base_bdevs": 2, 00:35:23.019 "num_base_bdevs_discovered": 1, 00:35:23.019 "num_base_bdevs_operational": 1, 00:35:23.019 "base_bdevs_list": [ 00:35:23.019 { 00:35:23.019 "name": null, 00:35:23.019 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:23.019 "is_configured": false, 00:35:23.019 "data_offset": 0, 00:35:23.019 "data_size": 7936 00:35:23.019 }, 00:35:23.019 { 00:35:23.019 "name": "BaseBdev2", 00:35:23.019 "uuid": "8d3cd44f-9051-537b-b56e-5bafca80b145", 00:35:23.019 "is_configured": true, 00:35:23.019 "data_offset": 256, 00:35:23.019 "data_size": 7936 00:35:23.019 } 00:35:23.019 ] 00:35:23.019 }' 00:35:23.019 13:55:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:35:23.019 13:55:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:35:23.019 13:55:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:35:23.019 13:55:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:35:23.019 13:55:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@784 -- # killprocess 85165 00:35:23.019 13:55:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@954 -- # '[' -z 85165 ']' 00:35:23.019 13:55:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@958 -- # kill -0 85165 00:35:23.020 13:55:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@959 -- # uname 00:35:23.020 13:55:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:35:23.020 13:55:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 85165 00:35:23.020 13:55:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:35:23.020 killing process with pid 85165 00:35:23.020 13:55:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:35:23.020 13:55:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@972 -- # echo 'killing process with pid 85165' 00:35:23.020 Received shutdown signal, test time was about 60.000000 seconds 00:35:23.020 00:35:23.020 Latency(us) 00:35:23.020 [2024-11-20T13:55:30.079Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:35:23.020 [2024-11-20T13:55:30.079Z] =================================================================================================================== 00:35:23.020 [2024-11-20T13:55:30.079Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:35:23.020 13:55:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@973 -- # kill 85165 00:35:23.020 13:55:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@978 -- # wait 85165 00:35:23.020 [2024-11-20 13:55:30.067101] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:35:23.020 [2024-11-20 13:55:30.067225] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:35:23.020 [2024-11-20 13:55:30.067276] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:35:23.020 [2024-11-20 13:55:30.067286] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:35:23.279 [2024-11-20 13:55:30.231639] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:35:23.845 13:55:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@786 -- # return 0 00:35:23.845 00:35:23.845 real 0m17.127s 00:35:23.845 user 0m21.624s 00:35:23.845 sys 0m2.009s 00:35:23.845 13:55:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@1130 -- # xtrace_disable 00:35:23.845 13:55:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:35:23.845 ************************************ 00:35:23.845 END TEST raid_rebuild_test_sb_md_separate 00:35:23.845 ************************************ 00:35:23.845 13:55:30 bdev_raid -- bdev/bdev_raid.sh@1010 -- # base_malloc_params='-m 32 -i' 00:35:23.845 13:55:30 bdev_raid -- bdev/bdev_raid.sh@1011 -- # run_test raid_state_function_test_sb_md_interleaved raid_state_function_test raid1 2 true 00:35:23.845 13:55:30 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:35:23.845 13:55:30 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:35:23.845 13:55:30 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:35:23.845 ************************************ 00:35:23.845 START TEST raid_state_function_test_sb_md_interleaved 00:35:23.845 ************************************ 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # local strip_size 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:35:23.845 Process raid pid: 85835 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@229 -- # raid_pid=85835 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 85835' 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@231 -- # waitforlisten 85835 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@835 -- # '[' -z 85835 ']' 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@840 -- # local max_retries=100 00:35:23.845 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@844 -- # xtrace_disable 00:35:23.845 13:55:30 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:24.104 [2024-11-20 13:55:30.951415] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:35:24.104 [2024-11-20 13:55:30.951543] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:35:24.104 [2024-11-20 13:55:31.110587] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:35:24.361 [2024-11-20 13:55:31.205754] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:35:24.361 [2024-11-20 13:55:31.325154] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:35:24.361 [2024-11-20 13:55:31.325186] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:35:24.926 13:55:31 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:35:24.926 13:55:31 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@868 -- # return 0 00:35:24.926 13:55:31 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:35:24.926 13:55:31 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:24.926 13:55:31 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:24.926 [2024-11-20 13:55:31.794012] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:35:24.926 [2024-11-20 13:55:31.794072] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:35:24.926 [2024-11-20 13:55:31.794085] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:35:24.926 [2024-11-20 13:55:31.794093] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:35:24.926 13:55:31 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:24.926 13:55:31 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:35:24.926 13:55:31 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:35:24.926 13:55:31 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:35:24.926 13:55:31 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:24.926 13:55:31 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:24.926 13:55:31 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:35:24.926 13:55:31 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:24.926 13:55:31 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:24.926 13:55:31 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:24.926 13:55:31 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:24.926 13:55:31 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:24.926 13:55:31 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:35:24.926 13:55:31 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:24.926 13:55:31 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:24.926 13:55:31 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:24.926 13:55:31 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:24.926 "name": "Existed_Raid", 00:35:24.926 "uuid": "b237751a-32d6-4586-bbbd-0f8df01b62bb", 00:35:24.926 "strip_size_kb": 0, 00:35:24.926 "state": "configuring", 00:35:24.926 "raid_level": "raid1", 00:35:24.926 "superblock": true, 00:35:24.926 "num_base_bdevs": 2, 00:35:24.926 "num_base_bdevs_discovered": 0, 00:35:24.926 "num_base_bdevs_operational": 2, 00:35:24.926 "base_bdevs_list": [ 00:35:24.926 { 00:35:24.926 "name": "BaseBdev1", 00:35:24.926 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:24.926 "is_configured": false, 00:35:24.926 "data_offset": 0, 00:35:24.926 "data_size": 0 00:35:24.926 }, 00:35:24.926 { 00:35:24.926 "name": "BaseBdev2", 00:35:24.926 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:24.926 "is_configured": false, 00:35:24.926 "data_offset": 0, 00:35:24.926 "data_size": 0 00:35:24.926 } 00:35:24.926 ] 00:35:24.926 }' 00:35:24.926 13:55:31 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:24.926 13:55:31 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:25.184 [2024-11-20 13:55:32.126031] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:35:25.184 [2024-11-20 13:55:32.126066] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:25.184 [2024-11-20 13:55:32.134019] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:35:25.184 [2024-11-20 13:55:32.134056] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:35:25.184 [2024-11-20 13:55:32.134064] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:35:25.184 [2024-11-20 13:55:32.134074] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev1 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:25.184 [2024-11-20 13:55:32.163386] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:35:25.184 BaseBdev1 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@905 -- # local i 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:25.184 [ 00:35:25.184 { 00:35:25.184 "name": "BaseBdev1", 00:35:25.184 "aliases": [ 00:35:25.184 "40c4864e-a1ac-454c-9ba2-2b0dd27bbefe" 00:35:25.184 ], 00:35:25.184 "product_name": "Malloc disk", 00:35:25.184 "block_size": 4128, 00:35:25.184 "num_blocks": 8192, 00:35:25.184 "uuid": "40c4864e-a1ac-454c-9ba2-2b0dd27bbefe", 00:35:25.184 "md_size": 32, 00:35:25.184 "md_interleave": true, 00:35:25.184 "dif_type": 0, 00:35:25.184 "assigned_rate_limits": { 00:35:25.184 "rw_ios_per_sec": 0, 00:35:25.184 "rw_mbytes_per_sec": 0, 00:35:25.184 "r_mbytes_per_sec": 0, 00:35:25.184 "w_mbytes_per_sec": 0 00:35:25.184 }, 00:35:25.184 "claimed": true, 00:35:25.184 "claim_type": "exclusive_write", 00:35:25.184 "zoned": false, 00:35:25.184 "supported_io_types": { 00:35:25.184 "read": true, 00:35:25.184 "write": true, 00:35:25.184 "unmap": true, 00:35:25.184 "flush": true, 00:35:25.184 "reset": true, 00:35:25.184 "nvme_admin": false, 00:35:25.184 "nvme_io": false, 00:35:25.184 "nvme_io_md": false, 00:35:25.184 "write_zeroes": true, 00:35:25.184 "zcopy": true, 00:35:25.184 "get_zone_info": false, 00:35:25.184 "zone_management": false, 00:35:25.184 "zone_append": false, 00:35:25.184 "compare": false, 00:35:25.184 "compare_and_write": false, 00:35:25.184 "abort": true, 00:35:25.184 "seek_hole": false, 00:35:25.184 "seek_data": false, 00:35:25.184 "copy": true, 00:35:25.184 "nvme_iov_md": false 00:35:25.184 }, 00:35:25.184 "memory_domains": [ 00:35:25.184 { 00:35:25.184 "dma_device_id": "system", 00:35:25.184 "dma_device_type": 1 00:35:25.184 }, 00:35:25.184 { 00:35:25.184 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:35:25.184 "dma_device_type": 2 00:35:25.184 } 00:35:25.184 ], 00:35:25.184 "driver_specific": {} 00:35:25.184 } 00:35:25.184 ] 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@911 -- # return 0 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:25.184 "name": "Existed_Raid", 00:35:25.184 "uuid": "12369bf8-1219-4ab2-b978-80d7f879ae9f", 00:35:25.184 "strip_size_kb": 0, 00:35:25.184 "state": "configuring", 00:35:25.184 "raid_level": "raid1", 00:35:25.184 "superblock": true, 00:35:25.184 "num_base_bdevs": 2, 00:35:25.184 "num_base_bdevs_discovered": 1, 00:35:25.184 "num_base_bdevs_operational": 2, 00:35:25.184 "base_bdevs_list": [ 00:35:25.184 { 00:35:25.184 "name": "BaseBdev1", 00:35:25.184 "uuid": "40c4864e-a1ac-454c-9ba2-2b0dd27bbefe", 00:35:25.184 "is_configured": true, 00:35:25.184 "data_offset": 256, 00:35:25.184 "data_size": 7936 00:35:25.184 }, 00:35:25.184 { 00:35:25.184 "name": "BaseBdev2", 00:35:25.184 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:25.184 "is_configured": false, 00:35:25.184 "data_offset": 0, 00:35:25.184 "data_size": 0 00:35:25.184 } 00:35:25.184 ] 00:35:25.184 }' 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:25.184 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:25.751 [2024-11-20 13:55:32.507502] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:35:25.751 [2024-11-20 13:55:32.507555] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:25.751 [2024-11-20 13:55:32.515544] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:35:25.751 [2024-11-20 13:55:32.517159] bdev.c:8666:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:35:25.751 [2024-11-20 13:55:32.517197] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:25.751 "name": "Existed_Raid", 00:35:25.751 "uuid": "8437ed2b-fc8f-41f7-a309-06d8343c2c93", 00:35:25.751 "strip_size_kb": 0, 00:35:25.751 "state": "configuring", 00:35:25.751 "raid_level": "raid1", 00:35:25.751 "superblock": true, 00:35:25.751 "num_base_bdevs": 2, 00:35:25.751 "num_base_bdevs_discovered": 1, 00:35:25.751 "num_base_bdevs_operational": 2, 00:35:25.751 "base_bdevs_list": [ 00:35:25.751 { 00:35:25.751 "name": "BaseBdev1", 00:35:25.751 "uuid": "40c4864e-a1ac-454c-9ba2-2b0dd27bbefe", 00:35:25.751 "is_configured": true, 00:35:25.751 "data_offset": 256, 00:35:25.751 "data_size": 7936 00:35:25.751 }, 00:35:25.751 { 00:35:25.751 "name": "BaseBdev2", 00:35:25.751 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:25.751 "is_configured": false, 00:35:25.751 "data_offset": 0, 00:35:25.751 "data_size": 0 00:35:25.751 } 00:35:25.751 ] 00:35:25.751 }' 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:25.751 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:26.010 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev2 00:35:26.010 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:26.010 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:26.010 [2024-11-20 13:55:32.859240] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:35:26.010 [2024-11-20 13:55:32.859405] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:35:26.010 [2024-11-20 13:55:32.859415] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:35:26.010 [2024-11-20 13:55:32.859483] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:35:26.010 [2024-11-20 13:55:32.859545] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:35:26.010 [2024-11-20 13:55:32.859590] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:35:26.010 [2024-11-20 13:55:32.859637] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:35:26.010 BaseBdev2 00:35:26.010 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:26.010 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:35:26.010 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:35:26.010 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:35:26.010 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@905 -- # local i 00:35:26.010 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:35:26.010 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:35:26.010 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:35:26.010 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:26.010 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:26.010 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:26.010 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:35:26.010 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:26.010 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:26.010 [ 00:35:26.010 { 00:35:26.010 "name": "BaseBdev2", 00:35:26.010 "aliases": [ 00:35:26.010 "b08c95ef-a69f-4ca8-9059-a294fb70c457" 00:35:26.010 ], 00:35:26.010 "product_name": "Malloc disk", 00:35:26.010 "block_size": 4128, 00:35:26.010 "num_blocks": 8192, 00:35:26.010 "uuid": "b08c95ef-a69f-4ca8-9059-a294fb70c457", 00:35:26.010 "md_size": 32, 00:35:26.010 "md_interleave": true, 00:35:26.010 "dif_type": 0, 00:35:26.010 "assigned_rate_limits": { 00:35:26.010 "rw_ios_per_sec": 0, 00:35:26.010 "rw_mbytes_per_sec": 0, 00:35:26.010 "r_mbytes_per_sec": 0, 00:35:26.010 "w_mbytes_per_sec": 0 00:35:26.010 }, 00:35:26.010 "claimed": true, 00:35:26.010 "claim_type": "exclusive_write", 00:35:26.010 "zoned": false, 00:35:26.010 "supported_io_types": { 00:35:26.010 "read": true, 00:35:26.010 "write": true, 00:35:26.010 "unmap": true, 00:35:26.010 "flush": true, 00:35:26.010 "reset": true, 00:35:26.010 "nvme_admin": false, 00:35:26.010 "nvme_io": false, 00:35:26.010 "nvme_io_md": false, 00:35:26.010 "write_zeroes": true, 00:35:26.010 "zcopy": true, 00:35:26.010 "get_zone_info": false, 00:35:26.010 "zone_management": false, 00:35:26.010 "zone_append": false, 00:35:26.010 "compare": false, 00:35:26.010 "compare_and_write": false, 00:35:26.010 "abort": true, 00:35:26.010 "seek_hole": false, 00:35:26.010 "seek_data": false, 00:35:26.010 "copy": true, 00:35:26.010 "nvme_iov_md": false 00:35:26.010 }, 00:35:26.010 "memory_domains": [ 00:35:26.010 { 00:35:26.010 "dma_device_id": "system", 00:35:26.010 "dma_device_type": 1 00:35:26.010 }, 00:35:26.010 { 00:35:26.010 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:35:26.010 "dma_device_type": 2 00:35:26.010 } 00:35:26.010 ], 00:35:26.010 "driver_specific": {} 00:35:26.010 } 00:35:26.010 ] 00:35:26.011 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:26.011 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@911 -- # return 0 00:35:26.011 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:35:26.011 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:35:26.011 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:35:26.011 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:35:26.011 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:26.011 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:26.011 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:26.011 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:35:26.011 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:26.011 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:26.011 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:26.011 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:26.011 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:26.011 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:26.011 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:35:26.011 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:26.011 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:26.011 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:26.011 "name": "Existed_Raid", 00:35:26.011 "uuid": "8437ed2b-fc8f-41f7-a309-06d8343c2c93", 00:35:26.011 "strip_size_kb": 0, 00:35:26.011 "state": "online", 00:35:26.011 "raid_level": "raid1", 00:35:26.011 "superblock": true, 00:35:26.011 "num_base_bdevs": 2, 00:35:26.011 "num_base_bdevs_discovered": 2, 00:35:26.011 "num_base_bdevs_operational": 2, 00:35:26.011 "base_bdevs_list": [ 00:35:26.011 { 00:35:26.011 "name": "BaseBdev1", 00:35:26.011 "uuid": "40c4864e-a1ac-454c-9ba2-2b0dd27bbefe", 00:35:26.011 "is_configured": true, 00:35:26.011 "data_offset": 256, 00:35:26.011 "data_size": 7936 00:35:26.011 }, 00:35:26.011 { 00:35:26.011 "name": "BaseBdev2", 00:35:26.011 "uuid": "b08c95ef-a69f-4ca8-9059-a294fb70c457", 00:35:26.011 "is_configured": true, 00:35:26.011 "data_offset": 256, 00:35:26.011 "data_size": 7936 00:35:26.011 } 00:35:26.011 ] 00:35:26.011 }' 00:35:26.011 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:26.011 13:55:32 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:26.269 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:35:26.269 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:35:26.269 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:35:26.269 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:35:26.269 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:35:26.269 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:35:26.269 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:35:26.269 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:35:26.269 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:26.269 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:26.269 [2024-11-20 13:55:33.219626] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:35:26.269 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:26.269 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:35:26.269 "name": "Existed_Raid", 00:35:26.269 "aliases": [ 00:35:26.269 "8437ed2b-fc8f-41f7-a309-06d8343c2c93" 00:35:26.269 ], 00:35:26.269 "product_name": "Raid Volume", 00:35:26.269 "block_size": 4128, 00:35:26.269 "num_blocks": 7936, 00:35:26.269 "uuid": "8437ed2b-fc8f-41f7-a309-06d8343c2c93", 00:35:26.269 "md_size": 32, 00:35:26.269 "md_interleave": true, 00:35:26.269 "dif_type": 0, 00:35:26.269 "assigned_rate_limits": { 00:35:26.269 "rw_ios_per_sec": 0, 00:35:26.269 "rw_mbytes_per_sec": 0, 00:35:26.269 "r_mbytes_per_sec": 0, 00:35:26.269 "w_mbytes_per_sec": 0 00:35:26.269 }, 00:35:26.269 "claimed": false, 00:35:26.269 "zoned": false, 00:35:26.269 "supported_io_types": { 00:35:26.269 "read": true, 00:35:26.269 "write": true, 00:35:26.269 "unmap": false, 00:35:26.269 "flush": false, 00:35:26.269 "reset": true, 00:35:26.269 "nvme_admin": false, 00:35:26.269 "nvme_io": false, 00:35:26.269 "nvme_io_md": false, 00:35:26.269 "write_zeroes": true, 00:35:26.269 "zcopy": false, 00:35:26.269 "get_zone_info": false, 00:35:26.269 "zone_management": false, 00:35:26.269 "zone_append": false, 00:35:26.269 "compare": false, 00:35:26.269 "compare_and_write": false, 00:35:26.269 "abort": false, 00:35:26.269 "seek_hole": false, 00:35:26.269 "seek_data": false, 00:35:26.269 "copy": false, 00:35:26.269 "nvme_iov_md": false 00:35:26.269 }, 00:35:26.269 "memory_domains": [ 00:35:26.269 { 00:35:26.269 "dma_device_id": "system", 00:35:26.269 "dma_device_type": 1 00:35:26.269 }, 00:35:26.269 { 00:35:26.269 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:35:26.269 "dma_device_type": 2 00:35:26.269 }, 00:35:26.269 { 00:35:26.269 "dma_device_id": "system", 00:35:26.269 "dma_device_type": 1 00:35:26.269 }, 00:35:26.269 { 00:35:26.269 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:35:26.269 "dma_device_type": 2 00:35:26.269 } 00:35:26.269 ], 00:35:26.269 "driver_specific": { 00:35:26.269 "raid": { 00:35:26.269 "uuid": "8437ed2b-fc8f-41f7-a309-06d8343c2c93", 00:35:26.269 "strip_size_kb": 0, 00:35:26.269 "state": "online", 00:35:26.269 "raid_level": "raid1", 00:35:26.269 "superblock": true, 00:35:26.269 "num_base_bdevs": 2, 00:35:26.269 "num_base_bdevs_discovered": 2, 00:35:26.269 "num_base_bdevs_operational": 2, 00:35:26.269 "base_bdevs_list": [ 00:35:26.269 { 00:35:26.269 "name": "BaseBdev1", 00:35:26.269 "uuid": "40c4864e-a1ac-454c-9ba2-2b0dd27bbefe", 00:35:26.269 "is_configured": true, 00:35:26.269 "data_offset": 256, 00:35:26.269 "data_size": 7936 00:35:26.269 }, 00:35:26.269 { 00:35:26.269 "name": "BaseBdev2", 00:35:26.269 "uuid": "b08c95ef-a69f-4ca8-9059-a294fb70c457", 00:35:26.269 "is_configured": true, 00:35:26.269 "data_offset": 256, 00:35:26.269 "data_size": 7936 00:35:26.269 } 00:35:26.269 ] 00:35:26.269 } 00:35:26.269 } 00:35:26.269 }' 00:35:26.269 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:35:26.269 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:35:26.269 BaseBdev2' 00:35:26.269 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:35:26.269 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:35:26.269 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:35:26.269 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:35:26.269 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:35:26.269 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:26.269 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:26.269 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:26.527 [2024-11-20 13:55:33.391418] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@260 -- # local expected_state 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@198 -- # case $1 in 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@199 -- # return 0 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:26.527 "name": "Existed_Raid", 00:35:26.527 "uuid": "8437ed2b-fc8f-41f7-a309-06d8343c2c93", 00:35:26.527 "strip_size_kb": 0, 00:35:26.527 "state": "online", 00:35:26.527 "raid_level": "raid1", 00:35:26.527 "superblock": true, 00:35:26.527 "num_base_bdevs": 2, 00:35:26.527 "num_base_bdevs_discovered": 1, 00:35:26.527 "num_base_bdevs_operational": 1, 00:35:26.527 "base_bdevs_list": [ 00:35:26.527 { 00:35:26.527 "name": null, 00:35:26.527 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:26.527 "is_configured": false, 00:35:26.527 "data_offset": 0, 00:35:26.527 "data_size": 7936 00:35:26.527 }, 00:35:26.527 { 00:35:26.527 "name": "BaseBdev2", 00:35:26.527 "uuid": "b08c95ef-a69f-4ca8-9059-a294fb70c457", 00:35:26.527 "is_configured": true, 00:35:26.527 "data_offset": 256, 00:35:26.527 "data_size": 7936 00:35:26.527 } 00:35:26.527 ] 00:35:26.527 }' 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:26.527 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:26.785 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:35:26.785 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:35:26.785 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:26.785 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:35:26.785 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:26.785 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:26.785 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:26.785 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:35:26.785 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:35:26.785 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:35:26.785 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:26.785 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:26.785 [2024-11-20 13:55:33.799758] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:35:26.785 [2024-11-20 13:55:33.799969] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:35:27.044 [2024-11-20 13:55:33.847851] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:35:27.044 [2024-11-20 13:55:33.847892] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:35:27.044 [2024-11-20 13:55:33.847903] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:35:27.044 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:27.044 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:35:27.044 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:35:27.044 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:27.044 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:35:27.044 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:27.044 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:27.044 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:27.044 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:35:27.044 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:35:27.044 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:35:27.044 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@326 -- # killprocess 85835 00:35:27.044 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@954 -- # '[' -z 85835 ']' 00:35:27.044 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@958 -- # kill -0 85835 00:35:27.044 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # uname 00:35:27.044 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:35:27.044 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 85835 00:35:27.044 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:35:27.044 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:35:27.044 killing process with pid 85835 00:35:27.044 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@972 -- # echo 'killing process with pid 85835' 00:35:27.044 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@973 -- # kill 85835 00:35:27.044 [2024-11-20 13:55:33.910939] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:35:27.044 13:55:33 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@978 -- # wait 85835 00:35:27.044 [2024-11-20 13:55:33.919671] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:35:27.611 13:55:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@328 -- # return 0 00:35:27.611 ************************************ 00:35:27.611 END TEST raid_state_function_test_sb_md_interleaved 00:35:27.611 ************************************ 00:35:27.611 00:35:27.611 real 0m3.646s 00:35:27.611 user 0m5.310s 00:35:27.611 sys 0m0.621s 00:35:27.611 13:55:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@1130 -- # xtrace_disable 00:35:27.611 13:55:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:27.611 13:55:34 bdev_raid -- bdev/bdev_raid.sh@1012 -- # run_test raid_superblock_test_md_interleaved raid_superblock_test raid1 2 00:35:27.611 13:55:34 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:35:27.611 13:55:34 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:35:27.611 13:55:34 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:35:27.611 ************************************ 00:35:27.611 START TEST raid_superblock_test_md_interleaved 00:35:27.611 ************************************ 00:35:27.611 13:55:34 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:35:27.611 13:55:34 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:35:27.611 13:55:34 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:35:27.611 13:55:34 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:35:27.611 13:55:34 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:35:27.611 13:55:34 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:35:27.611 13:55:34 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:35:27.611 13:55:34 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:35:27.611 13:55:34 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:35:27.611 13:55:34 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:35:27.611 13:55:34 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@399 -- # local strip_size 00:35:27.611 13:55:34 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:35:27.611 13:55:34 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:35:27.611 13:55:34 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:35:27.611 13:55:34 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:35:27.611 13:55:34 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:35:27.611 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:35:27.611 13:55:34 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@412 -- # raid_pid=86072 00:35:27.611 13:55:34 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@413 -- # waitforlisten 86072 00:35:27.611 13:55:34 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@835 -- # '[' -z 86072 ']' 00:35:27.611 13:55:34 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:35:27.611 13:55:34 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@840 -- # local max_retries=100 00:35:27.611 13:55:34 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:35:27.611 13:55:34 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@844 -- # xtrace_disable 00:35:27.611 13:55:34 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:27.611 13:55:34 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:35:27.611 [2024-11-20 13:55:34.641518] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:35:27.611 [2024-11-20 13:55:34.641788] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid86072 ] 00:35:27.868 [2024-11-20 13:55:34.795517] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:35:27.868 [2024-11-20 13:55:34.892757] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:35:28.127 [2024-11-20 13:55:35.010143] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:35:28.127 [2024-11-20 13:55:35.010191] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:35:28.694 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@868 -- # return 0 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b malloc1 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:28.695 malloc1 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:28.695 [2024-11-20 13:55:35.514759] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:35:28.695 [2024-11-20 13:55:35.514819] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:35:28.695 [2024-11-20 13:55:35.514839] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:35:28.695 [2024-11-20 13:55:35.514847] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:35:28.695 [2024-11-20 13:55:35.516521] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:35:28.695 [2024-11-20 13:55:35.516704] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:35:28.695 pt1 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b malloc2 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:28.695 malloc2 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:28.695 [2024-11-20 13:55:35.547667] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:35:28.695 [2024-11-20 13:55:35.547803] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:35:28.695 [2024-11-20 13:55:35.547824] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:35:28.695 [2024-11-20 13:55:35.547832] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:35:28.695 [2024-11-20 13:55:35.549433] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:35:28.695 [2024-11-20 13:55:35.549457] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:35:28.695 pt2 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:28.695 [2024-11-20 13:55:35.555702] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:35:28.695 [2024-11-20 13:55:35.557290] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:35:28.695 [2024-11-20 13:55:35.557441] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:35:28.695 [2024-11-20 13:55:35.557451] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:35:28.695 [2024-11-20 13:55:35.557514] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:35:28.695 [2024-11-20 13:55:35.557572] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:35:28.695 [2024-11-20 13:55:35.557581] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:35:28.695 [2024-11-20 13:55:35.557637] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:28.695 "name": "raid_bdev1", 00:35:28.695 "uuid": "bc904b60-d324-4729-8ae3-07ea4d3594b9", 00:35:28.695 "strip_size_kb": 0, 00:35:28.695 "state": "online", 00:35:28.695 "raid_level": "raid1", 00:35:28.695 "superblock": true, 00:35:28.695 "num_base_bdevs": 2, 00:35:28.695 "num_base_bdevs_discovered": 2, 00:35:28.695 "num_base_bdevs_operational": 2, 00:35:28.695 "base_bdevs_list": [ 00:35:28.695 { 00:35:28.695 "name": "pt1", 00:35:28.695 "uuid": "00000000-0000-0000-0000-000000000001", 00:35:28.695 "is_configured": true, 00:35:28.695 "data_offset": 256, 00:35:28.695 "data_size": 7936 00:35:28.695 }, 00:35:28.695 { 00:35:28.695 "name": "pt2", 00:35:28.695 "uuid": "00000000-0000-0000-0000-000000000002", 00:35:28.695 "is_configured": true, 00:35:28.695 "data_offset": 256, 00:35:28.695 "data_size": 7936 00:35:28.695 } 00:35:28.695 ] 00:35:28.695 }' 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:28.695 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:28.954 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:35:28.954 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:35:28.954 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:35:28.954 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:35:28.954 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:35:28.954 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:35:28.954 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:35:28.954 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:35:28.954 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:28.954 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:28.954 [2024-11-20 13:55:35.880049] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:35:28.954 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:28.954 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:35:28.954 "name": "raid_bdev1", 00:35:28.954 "aliases": [ 00:35:28.954 "bc904b60-d324-4729-8ae3-07ea4d3594b9" 00:35:28.954 ], 00:35:28.954 "product_name": "Raid Volume", 00:35:28.954 "block_size": 4128, 00:35:28.954 "num_blocks": 7936, 00:35:28.954 "uuid": "bc904b60-d324-4729-8ae3-07ea4d3594b9", 00:35:28.954 "md_size": 32, 00:35:28.954 "md_interleave": true, 00:35:28.954 "dif_type": 0, 00:35:28.954 "assigned_rate_limits": { 00:35:28.954 "rw_ios_per_sec": 0, 00:35:28.954 "rw_mbytes_per_sec": 0, 00:35:28.954 "r_mbytes_per_sec": 0, 00:35:28.954 "w_mbytes_per_sec": 0 00:35:28.954 }, 00:35:28.954 "claimed": false, 00:35:28.954 "zoned": false, 00:35:28.954 "supported_io_types": { 00:35:28.954 "read": true, 00:35:28.954 "write": true, 00:35:28.954 "unmap": false, 00:35:28.954 "flush": false, 00:35:28.954 "reset": true, 00:35:28.954 "nvme_admin": false, 00:35:28.954 "nvme_io": false, 00:35:28.954 "nvme_io_md": false, 00:35:28.954 "write_zeroes": true, 00:35:28.954 "zcopy": false, 00:35:28.954 "get_zone_info": false, 00:35:28.954 "zone_management": false, 00:35:28.954 "zone_append": false, 00:35:28.954 "compare": false, 00:35:28.954 "compare_and_write": false, 00:35:28.954 "abort": false, 00:35:28.954 "seek_hole": false, 00:35:28.954 "seek_data": false, 00:35:28.954 "copy": false, 00:35:28.954 "nvme_iov_md": false 00:35:28.954 }, 00:35:28.954 "memory_domains": [ 00:35:28.954 { 00:35:28.954 "dma_device_id": "system", 00:35:28.954 "dma_device_type": 1 00:35:28.954 }, 00:35:28.954 { 00:35:28.954 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:35:28.954 "dma_device_type": 2 00:35:28.954 }, 00:35:28.954 { 00:35:28.954 "dma_device_id": "system", 00:35:28.954 "dma_device_type": 1 00:35:28.954 }, 00:35:28.954 { 00:35:28.954 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:35:28.954 "dma_device_type": 2 00:35:28.954 } 00:35:28.954 ], 00:35:28.954 "driver_specific": { 00:35:28.954 "raid": { 00:35:28.954 "uuid": "bc904b60-d324-4729-8ae3-07ea4d3594b9", 00:35:28.954 "strip_size_kb": 0, 00:35:28.954 "state": "online", 00:35:28.954 "raid_level": "raid1", 00:35:28.954 "superblock": true, 00:35:28.954 "num_base_bdevs": 2, 00:35:28.954 "num_base_bdevs_discovered": 2, 00:35:28.954 "num_base_bdevs_operational": 2, 00:35:28.954 "base_bdevs_list": [ 00:35:28.954 { 00:35:28.954 "name": "pt1", 00:35:28.954 "uuid": "00000000-0000-0000-0000-000000000001", 00:35:28.954 "is_configured": true, 00:35:28.954 "data_offset": 256, 00:35:28.954 "data_size": 7936 00:35:28.954 }, 00:35:28.954 { 00:35:28.954 "name": "pt2", 00:35:28.954 "uuid": "00000000-0000-0000-0000-000000000002", 00:35:28.954 "is_configured": true, 00:35:28.954 "data_offset": 256, 00:35:28.954 "data_size": 7936 00:35:28.954 } 00:35:28.954 ] 00:35:28.954 } 00:35:28.954 } 00:35:28.954 }' 00:35:28.954 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:35:28.954 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:35:28.954 pt2' 00:35:28.954 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:35:28.954 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:35:28.954 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:35:28.955 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:35:28.955 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:35:28.955 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:28.955 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:28.955 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:28.955 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:35:28.955 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:35:28.955 13:55:35 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:35:28.955 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:35:28.955 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:28.955 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:28.955 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:35:28.955 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:29.214 [2024-11-20 13:55:36.044030] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=bc904b60-d324-4729-8ae3-07ea4d3594b9 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@436 -- # '[' -z bc904b60-d324-4729-8ae3-07ea4d3594b9 ']' 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:29.214 [2024-11-20 13:55:36.071760] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:35:29.214 [2024-11-20 13:55:36.071852] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:35:29.214 [2024-11-20 13:55:36.071968] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:35:29.214 [2024-11-20 13:55:36.072078] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:35:29.214 [2024-11-20 13:55:36.072136] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@652 -- # local es=0 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:29.214 [2024-11-20 13:55:36.163791] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:35:29.214 [2024-11-20 13:55:36.165436] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:35:29.214 [2024-11-20 13:55:36.165498] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:35:29.214 [2024-11-20 13:55:36.165544] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:35:29.214 [2024-11-20 13:55:36.165556] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:35:29.214 [2024-11-20 13:55:36.165565] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:35:29.214 request: 00:35:29.214 { 00:35:29.214 "name": "raid_bdev1", 00:35:29.214 "raid_level": "raid1", 00:35:29.214 "base_bdevs": [ 00:35:29.214 "malloc1", 00:35:29.214 "malloc2" 00:35:29.214 ], 00:35:29.214 "superblock": false, 00:35:29.214 "method": "bdev_raid_create", 00:35:29.214 "req_id": 1 00:35:29.214 } 00:35:29.214 Got JSON-RPC error response 00:35:29.214 response: 00:35:29.214 { 00:35:29.214 "code": -17, 00:35:29.214 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:35:29.214 } 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@655 -- # es=1 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:29.214 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:29.214 [2024-11-20 13:55:36.203775] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:35:29.214 [2024-11-20 13:55:36.203816] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:35:29.214 [2024-11-20 13:55:36.203830] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:35:29.215 [2024-11-20 13:55:36.203839] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:35:29.215 [2024-11-20 13:55:36.205499] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:35:29.215 [2024-11-20 13:55:36.205528] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:35:29.215 [2024-11-20 13:55:36.205569] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:35:29.215 [2024-11-20 13:55:36.205613] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:35:29.215 pt1 00:35:29.215 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:29.215 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:35:29.215 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:29.215 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:35:29.215 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:29.215 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:29.215 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:35:29.215 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:29.215 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:29.215 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:29.215 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:29.215 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:29.215 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:29.215 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:29.215 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:29.215 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:29.215 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:29.215 "name": "raid_bdev1", 00:35:29.215 "uuid": "bc904b60-d324-4729-8ae3-07ea4d3594b9", 00:35:29.215 "strip_size_kb": 0, 00:35:29.215 "state": "configuring", 00:35:29.215 "raid_level": "raid1", 00:35:29.215 "superblock": true, 00:35:29.215 "num_base_bdevs": 2, 00:35:29.215 "num_base_bdevs_discovered": 1, 00:35:29.215 "num_base_bdevs_operational": 2, 00:35:29.215 "base_bdevs_list": [ 00:35:29.215 { 00:35:29.215 "name": "pt1", 00:35:29.215 "uuid": "00000000-0000-0000-0000-000000000001", 00:35:29.215 "is_configured": true, 00:35:29.215 "data_offset": 256, 00:35:29.215 "data_size": 7936 00:35:29.215 }, 00:35:29.215 { 00:35:29.215 "name": null, 00:35:29.215 "uuid": "00000000-0000-0000-0000-000000000002", 00:35:29.215 "is_configured": false, 00:35:29.215 "data_offset": 256, 00:35:29.215 "data_size": 7936 00:35:29.215 } 00:35:29.215 ] 00:35:29.215 }' 00:35:29.215 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:29.215 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:29.474 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:35:29.474 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:35:29.474 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:35:29.474 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:35:29.474 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:29.474 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:29.474 [2024-11-20 13:55:36.511860] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:35:29.474 [2024-11-20 13:55:36.512075] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:35:29.474 [2024-11-20 13:55:36.512139] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:35:29.474 [2024-11-20 13:55:36.512463] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:35:29.474 [2024-11-20 13:55:36.512685] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:35:29.474 [2024-11-20 13:55:36.512768] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:35:29.474 [2024-11-20 13:55:36.512902] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:35:29.474 [2024-11-20 13:55:36.512942] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:35:29.474 [2024-11-20 13:55:36.513091] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:35:29.474 [2024-11-20 13:55:36.513151] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:35:29.474 [2024-11-20 13:55:36.513265] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:35:29.474 [2024-11-20 13:55:36.513370] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:35:29.474 [2024-11-20 13:55:36.513390] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:35:29.474 [2024-11-20 13:55:36.513503] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:35:29.474 pt2 00:35:29.474 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:29.474 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:35:29.474 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:35:29.474 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:35:29.474 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:29.474 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:29.474 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:29.474 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:29.474 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:35:29.474 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:29.474 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:29.474 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:29.474 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:29.474 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:29.474 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:29.474 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:29.474 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:29.474 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:29.732 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:29.732 "name": "raid_bdev1", 00:35:29.732 "uuid": "bc904b60-d324-4729-8ae3-07ea4d3594b9", 00:35:29.732 "strip_size_kb": 0, 00:35:29.732 "state": "online", 00:35:29.732 "raid_level": "raid1", 00:35:29.732 "superblock": true, 00:35:29.732 "num_base_bdevs": 2, 00:35:29.732 "num_base_bdevs_discovered": 2, 00:35:29.732 "num_base_bdevs_operational": 2, 00:35:29.732 "base_bdevs_list": [ 00:35:29.732 { 00:35:29.732 "name": "pt1", 00:35:29.732 "uuid": "00000000-0000-0000-0000-000000000001", 00:35:29.732 "is_configured": true, 00:35:29.732 "data_offset": 256, 00:35:29.732 "data_size": 7936 00:35:29.732 }, 00:35:29.732 { 00:35:29.732 "name": "pt2", 00:35:29.732 "uuid": "00000000-0000-0000-0000-000000000002", 00:35:29.732 "is_configured": true, 00:35:29.732 "data_offset": 256, 00:35:29.732 "data_size": 7936 00:35:29.732 } 00:35:29.732 ] 00:35:29.732 }' 00:35:29.732 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:29.732 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:29.991 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:35:29.991 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:35:29.991 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:35:29.991 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:35:29.991 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:35:29.991 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:35:29.991 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:35:29.991 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:35:29.991 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:29.991 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:29.991 [2024-11-20 13:55:36.836183] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:35:29.991 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:29.992 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:35:29.992 "name": "raid_bdev1", 00:35:29.992 "aliases": [ 00:35:29.992 "bc904b60-d324-4729-8ae3-07ea4d3594b9" 00:35:29.992 ], 00:35:29.992 "product_name": "Raid Volume", 00:35:29.992 "block_size": 4128, 00:35:29.992 "num_blocks": 7936, 00:35:29.992 "uuid": "bc904b60-d324-4729-8ae3-07ea4d3594b9", 00:35:29.992 "md_size": 32, 00:35:29.992 "md_interleave": true, 00:35:29.992 "dif_type": 0, 00:35:29.992 "assigned_rate_limits": { 00:35:29.992 "rw_ios_per_sec": 0, 00:35:29.992 "rw_mbytes_per_sec": 0, 00:35:29.992 "r_mbytes_per_sec": 0, 00:35:29.992 "w_mbytes_per_sec": 0 00:35:29.992 }, 00:35:29.992 "claimed": false, 00:35:29.992 "zoned": false, 00:35:29.992 "supported_io_types": { 00:35:29.992 "read": true, 00:35:29.992 "write": true, 00:35:29.992 "unmap": false, 00:35:29.992 "flush": false, 00:35:29.992 "reset": true, 00:35:29.992 "nvme_admin": false, 00:35:29.992 "nvme_io": false, 00:35:29.992 "nvme_io_md": false, 00:35:29.992 "write_zeroes": true, 00:35:29.992 "zcopy": false, 00:35:29.992 "get_zone_info": false, 00:35:29.992 "zone_management": false, 00:35:29.992 "zone_append": false, 00:35:29.992 "compare": false, 00:35:29.992 "compare_and_write": false, 00:35:29.992 "abort": false, 00:35:29.992 "seek_hole": false, 00:35:29.992 "seek_data": false, 00:35:29.992 "copy": false, 00:35:29.992 "nvme_iov_md": false 00:35:29.992 }, 00:35:29.992 "memory_domains": [ 00:35:29.992 { 00:35:29.992 "dma_device_id": "system", 00:35:29.992 "dma_device_type": 1 00:35:29.992 }, 00:35:29.992 { 00:35:29.992 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:35:29.992 "dma_device_type": 2 00:35:29.992 }, 00:35:29.992 { 00:35:29.992 "dma_device_id": "system", 00:35:29.992 "dma_device_type": 1 00:35:29.992 }, 00:35:29.992 { 00:35:29.992 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:35:29.992 "dma_device_type": 2 00:35:29.992 } 00:35:29.992 ], 00:35:29.992 "driver_specific": { 00:35:29.992 "raid": { 00:35:29.992 "uuid": "bc904b60-d324-4729-8ae3-07ea4d3594b9", 00:35:29.992 "strip_size_kb": 0, 00:35:29.992 "state": "online", 00:35:29.992 "raid_level": "raid1", 00:35:29.992 "superblock": true, 00:35:29.992 "num_base_bdevs": 2, 00:35:29.992 "num_base_bdevs_discovered": 2, 00:35:29.992 "num_base_bdevs_operational": 2, 00:35:29.992 "base_bdevs_list": [ 00:35:29.992 { 00:35:29.992 "name": "pt1", 00:35:29.992 "uuid": "00000000-0000-0000-0000-000000000001", 00:35:29.992 "is_configured": true, 00:35:29.992 "data_offset": 256, 00:35:29.992 "data_size": 7936 00:35:29.992 }, 00:35:29.992 { 00:35:29.992 "name": "pt2", 00:35:29.992 "uuid": "00000000-0000-0000-0000-000000000002", 00:35:29.992 "is_configured": true, 00:35:29.992 "data_offset": 256, 00:35:29.992 "data_size": 7936 00:35:29.992 } 00:35:29.992 ] 00:35:29.992 } 00:35:29.992 } 00:35:29.992 }' 00:35:29.992 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:35:29.992 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:35:29.992 pt2' 00:35:29.992 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:35:29.992 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:35:29.992 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:35:29.992 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:35:29.992 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:29.992 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:29.992 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:35:29.992 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:29.992 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:35:29.992 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:35:29.992 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:35:29.992 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:35:29.992 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:29.992 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:35:29.992 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:29.992 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:29.992 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:35:29.992 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:35:29.992 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:35:29.992 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:29.992 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:29.992 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:35:29.992 [2024-11-20 13:55:36.988163] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:35:29.992 13:55:36 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:29.992 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # '[' bc904b60-d324-4729-8ae3-07ea4d3594b9 '!=' bc904b60-d324-4729-8ae3-07ea4d3594b9 ']' 00:35:29.992 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:35:29.992 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@198 -- # case $1 in 00:35:29.992 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@199 -- # return 0 00:35:29.992 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:35:29.992 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:29.992 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:29.992 [2024-11-20 13:55:37.019955] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:35:29.992 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:29.992 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:35:29.992 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:29.992 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:29.992 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:29.992 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:29.992 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:35:29.993 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:29.993 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:29.993 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:29.993 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:29.993 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:29.993 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:29.993 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:29.993 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:29.993 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:30.250 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:30.250 "name": "raid_bdev1", 00:35:30.250 "uuid": "bc904b60-d324-4729-8ae3-07ea4d3594b9", 00:35:30.250 "strip_size_kb": 0, 00:35:30.250 "state": "online", 00:35:30.250 "raid_level": "raid1", 00:35:30.250 "superblock": true, 00:35:30.250 "num_base_bdevs": 2, 00:35:30.250 "num_base_bdevs_discovered": 1, 00:35:30.250 "num_base_bdevs_operational": 1, 00:35:30.250 "base_bdevs_list": [ 00:35:30.250 { 00:35:30.250 "name": null, 00:35:30.250 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:30.250 "is_configured": false, 00:35:30.250 "data_offset": 0, 00:35:30.250 "data_size": 7936 00:35:30.250 }, 00:35:30.250 { 00:35:30.250 "name": "pt2", 00:35:30.250 "uuid": "00000000-0000-0000-0000-000000000002", 00:35:30.250 "is_configured": true, 00:35:30.250 "data_offset": 256, 00:35:30.250 "data_size": 7936 00:35:30.250 } 00:35:30.250 ] 00:35:30.250 }' 00:35:30.250 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:30.250 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:30.557 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:30.558 [2024-11-20 13:55:37.327999] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:35:30.558 [2024-11-20 13:55:37.328021] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:35:30.558 [2024-11-20 13:55:37.328082] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:35:30.558 [2024-11-20 13:55:37.328129] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:35:30.558 [2024-11-20 13:55:37.328139] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@519 -- # i=1 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:30.558 [2024-11-20 13:55:37.380010] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:35:30.558 [2024-11-20 13:55:37.380122] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:35:30.558 [2024-11-20 13:55:37.380151] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:35:30.558 [2024-11-20 13:55:37.380191] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:35:30.558 [2024-11-20 13:55:37.381914] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:35:30.558 [2024-11-20 13:55:37.382013] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:35:30.558 [2024-11-20 13:55:37.382061] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:35:30.558 [2024-11-20 13:55:37.382105] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:35:30.558 [2024-11-20 13:55:37.382160] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:35:30.558 [2024-11-20 13:55:37.382170] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:35:30.558 [2024-11-20 13:55:37.382246] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:35:30.558 [2024-11-20 13:55:37.382296] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:35:30.558 [2024-11-20 13:55:37.382303] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:35:30.558 [2024-11-20 13:55:37.382351] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:35:30.558 pt2 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:30.558 "name": "raid_bdev1", 00:35:30.558 "uuid": "bc904b60-d324-4729-8ae3-07ea4d3594b9", 00:35:30.558 "strip_size_kb": 0, 00:35:30.558 "state": "online", 00:35:30.558 "raid_level": "raid1", 00:35:30.558 "superblock": true, 00:35:30.558 "num_base_bdevs": 2, 00:35:30.558 "num_base_bdevs_discovered": 1, 00:35:30.558 "num_base_bdevs_operational": 1, 00:35:30.558 "base_bdevs_list": [ 00:35:30.558 { 00:35:30.558 "name": null, 00:35:30.558 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:30.558 "is_configured": false, 00:35:30.558 "data_offset": 256, 00:35:30.558 "data_size": 7936 00:35:30.558 }, 00:35:30.558 { 00:35:30.558 "name": "pt2", 00:35:30.558 "uuid": "00000000-0000-0000-0000-000000000002", 00:35:30.558 "is_configured": true, 00:35:30.558 "data_offset": 256, 00:35:30.558 "data_size": 7936 00:35:30.558 } 00:35:30.558 ] 00:35:30.558 }' 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:30.558 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:30.816 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:30.817 [2024-11-20 13:55:37.708058] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:35:30.817 [2024-11-20 13:55:37.708083] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:35:30.817 [2024-11-20 13:55:37.708142] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:35:30.817 [2024-11-20 13:55:37.708190] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:35:30.817 [2024-11-20 13:55:37.708198] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:30.817 [2024-11-20 13:55:37.752077] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:35:30.817 [2024-11-20 13:55:37.752123] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:35:30.817 [2024-11-20 13:55:37.752139] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:35:30.817 [2024-11-20 13:55:37.752147] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:35:30.817 [2024-11-20 13:55:37.753876] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:35:30.817 [2024-11-20 13:55:37.753996] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:35:30.817 [2024-11-20 13:55:37.754052] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:35:30.817 [2024-11-20 13:55:37.754093] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:35:30.817 [2024-11-20 13:55:37.754181] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:35:30.817 [2024-11-20 13:55:37.754189] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:35:30.817 [2024-11-20 13:55:37.754203] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:35:30.817 [2024-11-20 13:55:37.754244] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:35:30.817 [2024-11-20 13:55:37.754304] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:35:30.817 [2024-11-20 13:55:37.754312] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:35:30.817 [2024-11-20 13:55:37.754374] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:35:30.817 [2024-11-20 13:55:37.754420] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:35:30.817 [2024-11-20 13:55:37.754429] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:35:30.817 [2024-11-20 13:55:37.754483] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:35:30.817 pt1 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:30.817 "name": "raid_bdev1", 00:35:30.817 "uuid": "bc904b60-d324-4729-8ae3-07ea4d3594b9", 00:35:30.817 "strip_size_kb": 0, 00:35:30.817 "state": "online", 00:35:30.817 "raid_level": "raid1", 00:35:30.817 "superblock": true, 00:35:30.817 "num_base_bdevs": 2, 00:35:30.817 "num_base_bdevs_discovered": 1, 00:35:30.817 "num_base_bdevs_operational": 1, 00:35:30.817 "base_bdevs_list": [ 00:35:30.817 { 00:35:30.817 "name": null, 00:35:30.817 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:30.817 "is_configured": false, 00:35:30.817 "data_offset": 256, 00:35:30.817 "data_size": 7936 00:35:30.817 }, 00:35:30.817 { 00:35:30.817 "name": "pt2", 00:35:30.817 "uuid": "00000000-0000-0000-0000-000000000002", 00:35:30.817 "is_configured": true, 00:35:30.817 "data_offset": 256, 00:35:30.817 "data_size": 7936 00:35:30.817 } 00:35:30.817 ] 00:35:30.817 }' 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:30.817 13:55:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:31.075 13:55:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:35:31.075 13:55:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:31.075 13:55:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:31.075 13:55:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:35:31.075 13:55:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:31.075 13:55:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:35:31.075 13:55:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:35:31.075 13:55:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:31.075 13:55:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:31.075 13:55:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:35:31.075 [2024-11-20 13:55:38.104386] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:35:31.075 13:55:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:31.075 13:55:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # '[' bc904b60-d324-4729-8ae3-07ea4d3594b9 '!=' bc904b60-d324-4729-8ae3-07ea4d3594b9 ']' 00:35:31.075 13:55:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@563 -- # killprocess 86072 00:35:31.075 13:55:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@954 -- # '[' -z 86072 ']' 00:35:31.075 13:55:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@958 -- # kill -0 86072 00:35:31.332 13:55:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@959 -- # uname 00:35:31.332 13:55:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:35:31.332 13:55:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 86072 00:35:31.332 killing process with pid 86072 00:35:31.332 13:55:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:35:31.332 13:55:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:35:31.332 13:55:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@972 -- # echo 'killing process with pid 86072' 00:35:31.332 13:55:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@973 -- # kill 86072 00:35:31.332 [2024-11-20 13:55:38.157522] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:35:31.332 13:55:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@978 -- # wait 86072 00:35:31.332 [2024-11-20 13:55:38.157613] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:35:31.332 [2024-11-20 13:55:38.157662] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:35:31.332 [2024-11-20 13:55:38.157675] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:35:31.332 [2024-11-20 13:55:38.263413] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:35:31.898 13:55:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@565 -- # return 0 00:35:31.898 00:35:31.898 real 0m4.291s 00:35:31.898 user 0m6.470s 00:35:31.898 sys 0m0.795s 00:35:31.898 13:55:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@1130 -- # xtrace_disable 00:35:31.898 ************************************ 00:35:31.898 END TEST raid_superblock_test_md_interleaved 00:35:31.898 ************************************ 00:35:31.898 13:55:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:31.898 13:55:38 bdev_raid -- bdev/bdev_raid.sh@1013 -- # run_test raid_rebuild_test_sb_md_interleaved raid_rebuild_test raid1 2 true false false 00:35:31.898 13:55:38 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:35:31.898 13:55:38 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:35:31.898 13:55:38 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:35:31.898 ************************************ 00:35:31.898 START TEST raid_rebuild_test_sb_md_interleaved 00:35:31.898 ************************************ 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false false 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@573 -- # local verify=false 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # local strip_size 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@577 -- # local create_arg 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@579 -- # local data_offset 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:35:31.898 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@597 -- # raid_pid=86378 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@598 -- # waitforlisten 86378 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@835 -- # '[' -z 86378 ']' 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@840 -- # local max_retries=100 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@844 -- # xtrace_disable 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:31.898 13:55:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:35:32.155 I/O size of 3145728 is greater than zero copy threshold (65536). 00:35:32.155 Zero copy mechanism will not be used. 00:35:32.155 [2024-11-20 13:55:38.978373] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:35:32.155 [2024-11-20 13:55:38.978491] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid86378 ] 00:35:32.155 [2024-11-20 13:55:39.131743] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:35:32.413 [2024-11-20 13:55:39.227058] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:35:32.413 [2024-11-20 13:55:39.345607] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:35:32.413 [2024-11-20 13:55:39.345792] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:35:32.980 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:35:32.980 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@868 -- # return 0 00:35:32.980 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:35:32.980 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev1_malloc 00:35:32.980 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:32.980 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:32.980 BaseBdev1_malloc 00:35:32.980 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:32.980 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:35:32.980 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:32.980 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:32.980 [2024-11-20 13:55:39.846317] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:35:32.980 [2024-11-20 13:55:39.846382] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:35:32.980 [2024-11-20 13:55:39.846402] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:35:32.980 [2024-11-20 13:55:39.846412] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:35:32.980 [2024-11-20 13:55:39.848040] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:35:32.981 [2024-11-20 13:55:39.848245] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:35:32.981 BaseBdev1 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev2_malloc 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:32.981 BaseBdev2_malloc 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:32.981 [2024-11-20 13:55:39.879316] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:35:32.981 [2024-11-20 13:55:39.879458] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:35:32.981 [2024-11-20 13:55:39.879478] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:35:32.981 [2024-11-20 13:55:39.879489] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:35:32.981 [2024-11-20 13:55:39.881086] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:35:32.981 [2024-11-20 13:55:39.881115] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:35:32.981 BaseBdev2 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b spare_malloc 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:32.981 spare_malloc 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:32.981 spare_delay 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:32.981 [2024-11-20 13:55:39.932017] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:35:32.981 [2024-11-20 13:55:39.932068] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:35:32.981 [2024-11-20 13:55:39.932085] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:35:32.981 [2024-11-20 13:55:39.932095] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:35:32.981 [2024-11-20 13:55:39.933721] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:35:32.981 [2024-11-20 13:55:39.933752] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:35:32.981 spare 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:32.981 [2024-11-20 13:55:39.940065] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:35:32.981 [2024-11-20 13:55:39.941651] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:35:32.981 [2024-11-20 13:55:39.941805] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:35:32.981 [2024-11-20 13:55:39.941815] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:35:32.981 [2024-11-20 13:55:39.941876] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:35:32.981 [2024-11-20 13:55:39.941933] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:35:32.981 [2024-11-20 13:55:39.941940] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:35:32.981 [2024-11-20 13:55:39.942006] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:32.981 "name": "raid_bdev1", 00:35:32.981 "uuid": "c58e332f-a9bb-468c-8c1d-e91086adebaa", 00:35:32.981 "strip_size_kb": 0, 00:35:32.981 "state": "online", 00:35:32.981 "raid_level": "raid1", 00:35:32.981 "superblock": true, 00:35:32.981 "num_base_bdevs": 2, 00:35:32.981 "num_base_bdevs_discovered": 2, 00:35:32.981 "num_base_bdevs_operational": 2, 00:35:32.981 "base_bdevs_list": [ 00:35:32.981 { 00:35:32.981 "name": "BaseBdev1", 00:35:32.981 "uuid": "1d7473ad-c220-5853-842f-4a286c0eb603", 00:35:32.981 "is_configured": true, 00:35:32.981 "data_offset": 256, 00:35:32.981 "data_size": 7936 00:35:32.981 }, 00:35:32.981 { 00:35:32.981 "name": "BaseBdev2", 00:35:32.981 "uuid": "33c46252-4999-5be1-b419-0adb4c459922", 00:35:32.981 "is_configured": true, 00:35:32.981 "data_offset": 256, 00:35:32.981 "data_size": 7936 00:35:32.981 } 00:35:32.981 ] 00:35:32.981 }' 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:32.981 13:55:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:33.239 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:35:33.239 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:35:33.239 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:33.239 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:33.239 [2024-11-20 13:55:40.264387] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:35:33.239 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:33.239 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:35:33.239 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:33.239 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:35:33.239 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:33.239 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:33.497 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:33.497 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:35:33.497 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:35:33.497 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@624 -- # '[' false = true ']' 00:35:33.497 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:35:33.497 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:33.497 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:33.497 [2024-11-20 13:55:40.332111] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:35:33.497 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:33.497 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:35:33.497 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:33.497 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:33.497 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:33.497 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:33.497 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:35:33.497 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:33.497 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:33.497 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:33.497 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:33.497 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:33.497 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:33.498 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:33.498 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:33.498 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:33.498 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:33.498 "name": "raid_bdev1", 00:35:33.498 "uuid": "c58e332f-a9bb-468c-8c1d-e91086adebaa", 00:35:33.498 "strip_size_kb": 0, 00:35:33.498 "state": "online", 00:35:33.498 "raid_level": "raid1", 00:35:33.498 "superblock": true, 00:35:33.498 "num_base_bdevs": 2, 00:35:33.498 "num_base_bdevs_discovered": 1, 00:35:33.498 "num_base_bdevs_operational": 1, 00:35:33.498 "base_bdevs_list": [ 00:35:33.498 { 00:35:33.498 "name": null, 00:35:33.498 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:33.498 "is_configured": false, 00:35:33.498 "data_offset": 0, 00:35:33.498 "data_size": 7936 00:35:33.498 }, 00:35:33.498 { 00:35:33.498 "name": "BaseBdev2", 00:35:33.498 "uuid": "33c46252-4999-5be1-b419-0adb4c459922", 00:35:33.498 "is_configured": true, 00:35:33.498 "data_offset": 256, 00:35:33.498 "data_size": 7936 00:35:33.498 } 00:35:33.498 ] 00:35:33.498 }' 00:35:33.498 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:33.498 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:33.757 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:35:33.757 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:33.757 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:33.757 [2024-11-20 13:55:40.660214] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:35:33.757 [2024-11-20 13:55:40.670215] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:35:33.757 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:33.757 13:55:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@647 -- # sleep 1 00:35:33.757 [2024-11-20 13:55:40.671997] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:35:34.692 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:35:34.692 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:35:34.692 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:35:34.692 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:35:34.692 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:35:34.692 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:34.692 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:34.692 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:34.692 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:34.692 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:34.692 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:35:34.692 "name": "raid_bdev1", 00:35:34.692 "uuid": "c58e332f-a9bb-468c-8c1d-e91086adebaa", 00:35:34.692 "strip_size_kb": 0, 00:35:34.692 "state": "online", 00:35:34.692 "raid_level": "raid1", 00:35:34.692 "superblock": true, 00:35:34.692 "num_base_bdevs": 2, 00:35:34.692 "num_base_bdevs_discovered": 2, 00:35:34.692 "num_base_bdevs_operational": 2, 00:35:34.692 "process": { 00:35:34.692 "type": "rebuild", 00:35:34.692 "target": "spare", 00:35:34.692 "progress": { 00:35:34.692 "blocks": 2560, 00:35:34.692 "percent": 32 00:35:34.692 } 00:35:34.692 }, 00:35:34.692 "base_bdevs_list": [ 00:35:34.692 { 00:35:34.692 "name": "spare", 00:35:34.692 "uuid": "53668cc0-1137-5c6d-846c-24c671b41d63", 00:35:34.692 "is_configured": true, 00:35:34.692 "data_offset": 256, 00:35:34.692 "data_size": 7936 00:35:34.692 }, 00:35:34.692 { 00:35:34.692 "name": "BaseBdev2", 00:35:34.692 "uuid": "33c46252-4999-5be1-b419-0adb4c459922", 00:35:34.692 "is_configured": true, 00:35:34.692 "data_offset": 256, 00:35:34.692 "data_size": 7936 00:35:34.692 } 00:35:34.692 ] 00:35:34.692 }' 00:35:34.692 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:35:34.692 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:35:34.692 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:35:34.951 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:35:34.951 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:35:34.951 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:34.951 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:34.951 [2024-11-20 13:55:41.778373] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:35:34.951 [2024-11-20 13:55:41.878548] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:35:34.951 [2024-11-20 13:55:41.878612] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:35:34.951 [2024-11-20 13:55:41.878626] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:35:34.951 [2024-11-20 13:55:41.878638] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:35:34.951 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:34.951 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:35:34.951 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:34.951 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:34.951 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:34.951 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:34.951 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:35:34.951 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:34.951 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:34.951 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:34.951 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:34.951 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:34.951 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:34.951 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:34.951 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:34.951 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:34.951 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:34.951 "name": "raid_bdev1", 00:35:34.951 "uuid": "c58e332f-a9bb-468c-8c1d-e91086adebaa", 00:35:34.951 "strip_size_kb": 0, 00:35:34.951 "state": "online", 00:35:34.951 "raid_level": "raid1", 00:35:34.951 "superblock": true, 00:35:34.951 "num_base_bdevs": 2, 00:35:34.951 "num_base_bdevs_discovered": 1, 00:35:34.951 "num_base_bdevs_operational": 1, 00:35:34.951 "base_bdevs_list": [ 00:35:34.951 { 00:35:34.951 "name": null, 00:35:34.951 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:34.951 "is_configured": false, 00:35:34.951 "data_offset": 0, 00:35:34.951 "data_size": 7936 00:35:34.951 }, 00:35:34.951 { 00:35:34.951 "name": "BaseBdev2", 00:35:34.951 "uuid": "33c46252-4999-5be1-b419-0adb4c459922", 00:35:34.951 "is_configured": true, 00:35:34.951 "data_offset": 256, 00:35:34.951 "data_size": 7936 00:35:34.951 } 00:35:34.951 ] 00:35:34.951 }' 00:35:34.951 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:34.951 13:55:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:35.210 13:55:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:35:35.210 13:55:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:35:35.210 13:55:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:35:35.210 13:55:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:35:35.210 13:55:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:35:35.210 13:55:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:35.210 13:55:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:35.210 13:55:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:35.210 13:55:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:35.210 13:55:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:35.210 13:55:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:35:35.210 "name": "raid_bdev1", 00:35:35.210 "uuid": "c58e332f-a9bb-468c-8c1d-e91086adebaa", 00:35:35.210 "strip_size_kb": 0, 00:35:35.210 "state": "online", 00:35:35.210 "raid_level": "raid1", 00:35:35.210 "superblock": true, 00:35:35.210 "num_base_bdevs": 2, 00:35:35.210 "num_base_bdevs_discovered": 1, 00:35:35.210 "num_base_bdevs_operational": 1, 00:35:35.210 "base_bdevs_list": [ 00:35:35.210 { 00:35:35.210 "name": null, 00:35:35.210 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:35.210 "is_configured": false, 00:35:35.210 "data_offset": 0, 00:35:35.210 "data_size": 7936 00:35:35.210 }, 00:35:35.210 { 00:35:35.210 "name": "BaseBdev2", 00:35:35.210 "uuid": "33c46252-4999-5be1-b419-0adb4c459922", 00:35:35.210 "is_configured": true, 00:35:35.210 "data_offset": 256, 00:35:35.210 "data_size": 7936 00:35:35.210 } 00:35:35.210 ] 00:35:35.210 }' 00:35:35.210 13:55:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:35:35.469 13:55:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:35:35.469 13:55:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:35:35.469 13:55:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:35:35.469 13:55:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:35:35.469 13:55:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:35.469 13:55:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:35.469 [2024-11-20 13:55:42.326258] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:35:35.469 [2024-11-20 13:55:42.335627] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:35:35.469 13:55:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:35.469 13:55:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@663 -- # sleep 1 00:35:35.469 [2024-11-20 13:55:42.337275] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:35:36.403 "name": "raid_bdev1", 00:35:36.403 "uuid": "c58e332f-a9bb-468c-8c1d-e91086adebaa", 00:35:36.403 "strip_size_kb": 0, 00:35:36.403 "state": "online", 00:35:36.403 "raid_level": "raid1", 00:35:36.403 "superblock": true, 00:35:36.403 "num_base_bdevs": 2, 00:35:36.403 "num_base_bdevs_discovered": 2, 00:35:36.403 "num_base_bdevs_operational": 2, 00:35:36.403 "process": { 00:35:36.403 "type": "rebuild", 00:35:36.403 "target": "spare", 00:35:36.403 "progress": { 00:35:36.403 "blocks": 2560, 00:35:36.403 "percent": 32 00:35:36.403 } 00:35:36.403 }, 00:35:36.403 "base_bdevs_list": [ 00:35:36.403 { 00:35:36.403 "name": "spare", 00:35:36.403 "uuid": "53668cc0-1137-5c6d-846c-24c671b41d63", 00:35:36.403 "is_configured": true, 00:35:36.403 "data_offset": 256, 00:35:36.403 "data_size": 7936 00:35:36.403 }, 00:35:36.403 { 00:35:36.403 "name": "BaseBdev2", 00:35:36.403 "uuid": "33c46252-4999-5be1-b419-0adb4c459922", 00:35:36.403 "is_configured": true, 00:35:36.403 "data_offset": 256, 00:35:36.403 "data_size": 7936 00:35:36.403 } 00:35:36.403 ] 00:35:36.403 }' 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:35:36.403 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@706 -- # local timeout=590 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:36.403 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:36.662 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:35:36.662 "name": "raid_bdev1", 00:35:36.662 "uuid": "c58e332f-a9bb-468c-8c1d-e91086adebaa", 00:35:36.662 "strip_size_kb": 0, 00:35:36.662 "state": "online", 00:35:36.662 "raid_level": "raid1", 00:35:36.662 "superblock": true, 00:35:36.662 "num_base_bdevs": 2, 00:35:36.662 "num_base_bdevs_discovered": 2, 00:35:36.662 "num_base_bdevs_operational": 2, 00:35:36.662 "process": { 00:35:36.662 "type": "rebuild", 00:35:36.662 "target": "spare", 00:35:36.662 "progress": { 00:35:36.662 "blocks": 2816, 00:35:36.662 "percent": 35 00:35:36.662 } 00:35:36.662 }, 00:35:36.662 "base_bdevs_list": [ 00:35:36.662 { 00:35:36.662 "name": "spare", 00:35:36.662 "uuid": "53668cc0-1137-5c6d-846c-24c671b41d63", 00:35:36.662 "is_configured": true, 00:35:36.662 "data_offset": 256, 00:35:36.662 "data_size": 7936 00:35:36.662 }, 00:35:36.662 { 00:35:36.662 "name": "BaseBdev2", 00:35:36.662 "uuid": "33c46252-4999-5be1-b419-0adb4c459922", 00:35:36.662 "is_configured": true, 00:35:36.662 "data_offset": 256, 00:35:36.662 "data_size": 7936 00:35:36.662 } 00:35:36.662 ] 00:35:36.662 }' 00:35:36.662 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:35:36.662 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:35:36.662 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:35:36.662 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:35:36.662 13:55:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@711 -- # sleep 1 00:35:37.597 13:55:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:35:37.597 13:55:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:35:37.597 13:55:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:35:37.597 13:55:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:35:37.597 13:55:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:35:37.597 13:55:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:35:37.597 13:55:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:37.597 13:55:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:37.597 13:55:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:37.597 13:55:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:37.597 13:55:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:37.597 13:55:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:35:37.597 "name": "raid_bdev1", 00:35:37.597 "uuid": "c58e332f-a9bb-468c-8c1d-e91086adebaa", 00:35:37.597 "strip_size_kb": 0, 00:35:37.597 "state": "online", 00:35:37.597 "raid_level": "raid1", 00:35:37.597 "superblock": true, 00:35:37.597 "num_base_bdevs": 2, 00:35:37.597 "num_base_bdevs_discovered": 2, 00:35:37.597 "num_base_bdevs_operational": 2, 00:35:37.597 "process": { 00:35:37.597 "type": "rebuild", 00:35:37.597 "target": "spare", 00:35:37.597 "progress": { 00:35:37.597 "blocks": 5376, 00:35:37.597 "percent": 67 00:35:37.597 } 00:35:37.597 }, 00:35:37.597 "base_bdevs_list": [ 00:35:37.597 { 00:35:37.597 "name": "spare", 00:35:37.597 "uuid": "53668cc0-1137-5c6d-846c-24c671b41d63", 00:35:37.597 "is_configured": true, 00:35:37.597 "data_offset": 256, 00:35:37.597 "data_size": 7936 00:35:37.597 }, 00:35:37.597 { 00:35:37.597 "name": "BaseBdev2", 00:35:37.597 "uuid": "33c46252-4999-5be1-b419-0adb4c459922", 00:35:37.597 "is_configured": true, 00:35:37.597 "data_offset": 256, 00:35:37.597 "data_size": 7936 00:35:37.597 } 00:35:37.597 ] 00:35:37.597 }' 00:35:37.597 13:55:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:35:37.597 13:55:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:35:37.597 13:55:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:35:37.597 13:55:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:35:37.597 13:55:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@711 -- # sleep 1 00:35:38.529 [2024-11-20 13:55:45.454239] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:35:38.529 [2024-11-20 13:55:45.454311] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:35:38.529 [2024-11-20 13:55:45.454411] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:35:38.787 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:35:38.787 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:35:38.787 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:35:38.787 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:35:38.787 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:35:38.787 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:35:38.787 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:38.787 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:38.787 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:38.787 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:38.787 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:38.787 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:35:38.787 "name": "raid_bdev1", 00:35:38.787 "uuid": "c58e332f-a9bb-468c-8c1d-e91086adebaa", 00:35:38.787 "strip_size_kb": 0, 00:35:38.787 "state": "online", 00:35:38.787 "raid_level": "raid1", 00:35:38.787 "superblock": true, 00:35:38.787 "num_base_bdevs": 2, 00:35:38.787 "num_base_bdevs_discovered": 2, 00:35:38.787 "num_base_bdevs_operational": 2, 00:35:38.787 "base_bdevs_list": [ 00:35:38.787 { 00:35:38.787 "name": "spare", 00:35:38.787 "uuid": "53668cc0-1137-5c6d-846c-24c671b41d63", 00:35:38.787 "is_configured": true, 00:35:38.787 "data_offset": 256, 00:35:38.787 "data_size": 7936 00:35:38.787 }, 00:35:38.787 { 00:35:38.787 "name": "BaseBdev2", 00:35:38.787 "uuid": "33c46252-4999-5be1-b419-0adb4c459922", 00:35:38.787 "is_configured": true, 00:35:38.787 "data_offset": 256, 00:35:38.787 "data_size": 7936 00:35:38.787 } 00:35:38.787 ] 00:35:38.787 }' 00:35:38.787 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:35:38.787 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:35:38.787 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:35:38.787 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:35:38.787 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@709 -- # break 00:35:38.787 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:35:38.787 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:35:38.787 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:35:38.787 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:35:38.788 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:35:38.788 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:38.788 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:38.788 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:38.788 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:38.788 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:38.788 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:35:38.788 "name": "raid_bdev1", 00:35:38.788 "uuid": "c58e332f-a9bb-468c-8c1d-e91086adebaa", 00:35:38.788 "strip_size_kb": 0, 00:35:38.788 "state": "online", 00:35:38.788 "raid_level": "raid1", 00:35:38.788 "superblock": true, 00:35:38.788 "num_base_bdevs": 2, 00:35:38.788 "num_base_bdevs_discovered": 2, 00:35:38.788 "num_base_bdevs_operational": 2, 00:35:38.788 "base_bdevs_list": [ 00:35:38.788 { 00:35:38.788 "name": "spare", 00:35:38.788 "uuid": "53668cc0-1137-5c6d-846c-24c671b41d63", 00:35:38.788 "is_configured": true, 00:35:38.788 "data_offset": 256, 00:35:38.788 "data_size": 7936 00:35:38.788 }, 00:35:38.788 { 00:35:38.788 "name": "BaseBdev2", 00:35:38.788 "uuid": "33c46252-4999-5be1-b419-0adb4c459922", 00:35:38.788 "is_configured": true, 00:35:38.788 "data_offset": 256, 00:35:38.788 "data_size": 7936 00:35:38.788 } 00:35:38.788 ] 00:35:38.788 }' 00:35:38.788 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:35:38.788 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:35:38.788 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:35:38.788 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:35:38.788 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:35:38.788 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:38.788 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:38.788 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:38.788 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:38.788 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:35:38.788 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:38.788 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:38.788 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:38.788 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:38.788 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:38.788 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:38.788 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:38.788 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:39.046 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:39.046 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:39.046 "name": "raid_bdev1", 00:35:39.046 "uuid": "c58e332f-a9bb-468c-8c1d-e91086adebaa", 00:35:39.046 "strip_size_kb": 0, 00:35:39.046 "state": "online", 00:35:39.046 "raid_level": "raid1", 00:35:39.046 "superblock": true, 00:35:39.046 "num_base_bdevs": 2, 00:35:39.046 "num_base_bdevs_discovered": 2, 00:35:39.046 "num_base_bdevs_operational": 2, 00:35:39.046 "base_bdevs_list": [ 00:35:39.046 { 00:35:39.046 "name": "spare", 00:35:39.046 "uuid": "53668cc0-1137-5c6d-846c-24c671b41d63", 00:35:39.046 "is_configured": true, 00:35:39.046 "data_offset": 256, 00:35:39.046 "data_size": 7936 00:35:39.046 }, 00:35:39.046 { 00:35:39.046 "name": "BaseBdev2", 00:35:39.046 "uuid": "33c46252-4999-5be1-b419-0adb4c459922", 00:35:39.046 "is_configured": true, 00:35:39.046 "data_offset": 256, 00:35:39.046 "data_size": 7936 00:35:39.046 } 00:35:39.046 ] 00:35:39.046 }' 00:35:39.046 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:39.046 13:55:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:39.305 [2024-11-20 13:55:46.157930] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:35:39.305 [2024-11-20 13:55:46.157961] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:35:39.305 [2024-11-20 13:55:46.158063] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:35:39.305 [2024-11-20 13:55:46.158134] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:35:39.305 [2024-11-20 13:55:46.158144] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # jq length 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@722 -- # '[' false = true ']' 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:39.305 [2024-11-20 13:55:46.205920] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:35:39.305 [2024-11-20 13:55:46.206108] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:35:39.305 [2024-11-20 13:55:46.206134] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:35:39.305 [2024-11-20 13:55:46.206142] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:35:39.305 [2024-11-20 13:55:46.207907] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:35:39.305 [2024-11-20 13:55:46.207939] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:35:39.305 [2024-11-20 13:55:46.208000] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:35:39.305 [2024-11-20 13:55:46.208043] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:35:39.305 [2024-11-20 13:55:46.208132] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:35:39.305 spare 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:39.305 [2024-11-20 13:55:46.308209] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:35:39.305 [2024-11-20 13:55:46.308233] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:35:39.305 [2024-11-20 13:55:46.308323] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:35:39.305 [2024-11-20 13:55:46.308410] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:35:39.305 [2024-11-20 13:55:46.308418] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:35:39.305 [2024-11-20 13:55:46.308498] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:39.305 "name": "raid_bdev1", 00:35:39.305 "uuid": "c58e332f-a9bb-468c-8c1d-e91086adebaa", 00:35:39.305 "strip_size_kb": 0, 00:35:39.305 "state": "online", 00:35:39.305 "raid_level": "raid1", 00:35:39.305 "superblock": true, 00:35:39.305 "num_base_bdevs": 2, 00:35:39.305 "num_base_bdevs_discovered": 2, 00:35:39.305 "num_base_bdevs_operational": 2, 00:35:39.305 "base_bdevs_list": [ 00:35:39.305 { 00:35:39.305 "name": "spare", 00:35:39.305 "uuid": "53668cc0-1137-5c6d-846c-24c671b41d63", 00:35:39.305 "is_configured": true, 00:35:39.305 "data_offset": 256, 00:35:39.305 "data_size": 7936 00:35:39.305 }, 00:35:39.305 { 00:35:39.305 "name": "BaseBdev2", 00:35:39.305 "uuid": "33c46252-4999-5be1-b419-0adb4c459922", 00:35:39.305 "is_configured": true, 00:35:39.305 "data_offset": 256, 00:35:39.305 "data_size": 7936 00:35:39.305 } 00:35:39.305 ] 00:35:39.305 }' 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:39.305 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:39.564 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:35:39.564 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:35:39.564 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:35:39.564 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:35:39.564 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:35:39.564 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:39.564 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:39.564 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:39.564 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:35:39.822 "name": "raid_bdev1", 00:35:39.822 "uuid": "c58e332f-a9bb-468c-8c1d-e91086adebaa", 00:35:39.822 "strip_size_kb": 0, 00:35:39.822 "state": "online", 00:35:39.822 "raid_level": "raid1", 00:35:39.822 "superblock": true, 00:35:39.822 "num_base_bdevs": 2, 00:35:39.822 "num_base_bdevs_discovered": 2, 00:35:39.822 "num_base_bdevs_operational": 2, 00:35:39.822 "base_bdevs_list": [ 00:35:39.822 { 00:35:39.822 "name": "spare", 00:35:39.822 "uuid": "53668cc0-1137-5c6d-846c-24c671b41d63", 00:35:39.822 "is_configured": true, 00:35:39.822 "data_offset": 256, 00:35:39.822 "data_size": 7936 00:35:39.822 }, 00:35:39.822 { 00:35:39.822 "name": "BaseBdev2", 00:35:39.822 "uuid": "33c46252-4999-5be1-b419-0adb4c459922", 00:35:39.822 "is_configured": true, 00:35:39.822 "data_offset": 256, 00:35:39.822 "data_size": 7936 00:35:39.822 } 00:35:39.822 ] 00:35:39.822 }' 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:39.822 [2024-11-20 13:55:46.750083] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:39.822 "name": "raid_bdev1", 00:35:39.822 "uuid": "c58e332f-a9bb-468c-8c1d-e91086adebaa", 00:35:39.822 "strip_size_kb": 0, 00:35:39.822 "state": "online", 00:35:39.822 "raid_level": "raid1", 00:35:39.822 "superblock": true, 00:35:39.822 "num_base_bdevs": 2, 00:35:39.822 "num_base_bdevs_discovered": 1, 00:35:39.822 "num_base_bdevs_operational": 1, 00:35:39.822 "base_bdevs_list": [ 00:35:39.822 { 00:35:39.822 "name": null, 00:35:39.822 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:39.822 "is_configured": false, 00:35:39.822 "data_offset": 0, 00:35:39.822 "data_size": 7936 00:35:39.822 }, 00:35:39.822 { 00:35:39.822 "name": "BaseBdev2", 00:35:39.822 "uuid": "33c46252-4999-5be1-b419-0adb4c459922", 00:35:39.822 "is_configured": true, 00:35:39.822 "data_offset": 256, 00:35:39.822 "data_size": 7936 00:35:39.822 } 00:35:39.822 ] 00:35:39.822 }' 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:39.822 13:55:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:40.080 13:55:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:35:40.080 13:55:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:40.080 13:55:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:40.080 [2024-11-20 13:55:47.078155] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:35:40.080 [2024-11-20 13:55:47.078328] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:35:40.080 [2024-11-20 13:55:47.078342] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:35:40.080 [2024-11-20 13:55:47.078380] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:35:40.080 [2024-11-20 13:55:47.087854] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:35:40.080 13:55:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:40.080 13:55:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@757 -- # sleep 1 00:35:40.080 [2024-11-20 13:55:47.089586] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:35:41.518 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:35:41.518 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:35:41.518 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:35:41.518 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:35:41.518 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:35:41.518 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:41.518 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:41.518 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:41.518 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:41.518 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:41.518 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:35:41.518 "name": "raid_bdev1", 00:35:41.518 "uuid": "c58e332f-a9bb-468c-8c1d-e91086adebaa", 00:35:41.518 "strip_size_kb": 0, 00:35:41.518 "state": "online", 00:35:41.518 "raid_level": "raid1", 00:35:41.518 "superblock": true, 00:35:41.518 "num_base_bdevs": 2, 00:35:41.518 "num_base_bdevs_discovered": 2, 00:35:41.518 "num_base_bdevs_operational": 2, 00:35:41.518 "process": { 00:35:41.518 "type": "rebuild", 00:35:41.518 "target": "spare", 00:35:41.518 "progress": { 00:35:41.518 "blocks": 2560, 00:35:41.518 "percent": 32 00:35:41.518 } 00:35:41.518 }, 00:35:41.518 "base_bdevs_list": [ 00:35:41.518 { 00:35:41.518 "name": "spare", 00:35:41.518 "uuid": "53668cc0-1137-5c6d-846c-24c671b41d63", 00:35:41.518 "is_configured": true, 00:35:41.518 "data_offset": 256, 00:35:41.518 "data_size": 7936 00:35:41.518 }, 00:35:41.518 { 00:35:41.518 "name": "BaseBdev2", 00:35:41.518 "uuid": "33c46252-4999-5be1-b419-0adb4c459922", 00:35:41.518 "is_configured": true, 00:35:41.518 "data_offset": 256, 00:35:41.518 "data_size": 7936 00:35:41.518 } 00:35:41.518 ] 00:35:41.518 }' 00:35:41.518 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:35:41.518 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:41.519 [2024-11-20 13:55:48.191802] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:35:41.519 [2024-11-20 13:55:48.195411] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:35:41.519 [2024-11-20 13:55:48.195468] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:35:41.519 [2024-11-20 13:55:48.195481] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:35:41.519 [2024-11-20 13:55:48.195491] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:41.519 "name": "raid_bdev1", 00:35:41.519 "uuid": "c58e332f-a9bb-468c-8c1d-e91086adebaa", 00:35:41.519 "strip_size_kb": 0, 00:35:41.519 "state": "online", 00:35:41.519 "raid_level": "raid1", 00:35:41.519 "superblock": true, 00:35:41.519 "num_base_bdevs": 2, 00:35:41.519 "num_base_bdevs_discovered": 1, 00:35:41.519 "num_base_bdevs_operational": 1, 00:35:41.519 "base_bdevs_list": [ 00:35:41.519 { 00:35:41.519 "name": null, 00:35:41.519 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:41.519 "is_configured": false, 00:35:41.519 "data_offset": 0, 00:35:41.519 "data_size": 7936 00:35:41.519 }, 00:35:41.519 { 00:35:41.519 "name": "BaseBdev2", 00:35:41.519 "uuid": "33c46252-4999-5be1-b419-0adb4c459922", 00:35:41.519 "is_configured": true, 00:35:41.519 "data_offset": 256, 00:35:41.519 "data_size": 7936 00:35:41.519 } 00:35:41.519 ] 00:35:41.519 }' 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:41.519 [2024-11-20 13:55:48.531018] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:35:41.519 [2024-11-20 13:55:48.531086] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:35:41.519 [2024-11-20 13:55:48.531110] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:35:41.519 [2024-11-20 13:55:48.531120] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:35:41.519 [2024-11-20 13:55:48.531308] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:35:41.519 [2024-11-20 13:55:48.531320] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:35:41.519 [2024-11-20 13:55:48.531371] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:35:41.519 [2024-11-20 13:55:48.531383] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:35:41.519 [2024-11-20 13:55:48.531392] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:35:41.519 [2024-11-20 13:55:48.531408] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:35:41.519 [2024-11-20 13:55:48.540836] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:35:41.519 spare 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:41.519 13:55:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@764 -- # sleep 1 00:35:41.519 [2024-11-20 13:55:48.542512] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:35:42.910 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:35:42.910 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:35:42.910 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:35:42.910 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:35:42.910 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:35:42.910 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:42.910 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:42.910 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:42.910 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:42.910 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:42.910 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:35:42.910 "name": "raid_bdev1", 00:35:42.910 "uuid": "c58e332f-a9bb-468c-8c1d-e91086adebaa", 00:35:42.910 "strip_size_kb": 0, 00:35:42.910 "state": "online", 00:35:42.910 "raid_level": "raid1", 00:35:42.910 "superblock": true, 00:35:42.910 "num_base_bdevs": 2, 00:35:42.910 "num_base_bdevs_discovered": 2, 00:35:42.910 "num_base_bdevs_operational": 2, 00:35:42.910 "process": { 00:35:42.910 "type": "rebuild", 00:35:42.910 "target": "spare", 00:35:42.910 "progress": { 00:35:42.910 "blocks": 2560, 00:35:42.910 "percent": 32 00:35:42.910 } 00:35:42.910 }, 00:35:42.910 "base_bdevs_list": [ 00:35:42.910 { 00:35:42.910 "name": "spare", 00:35:42.910 "uuid": "53668cc0-1137-5c6d-846c-24c671b41d63", 00:35:42.910 "is_configured": true, 00:35:42.910 "data_offset": 256, 00:35:42.910 "data_size": 7936 00:35:42.910 }, 00:35:42.910 { 00:35:42.910 "name": "BaseBdev2", 00:35:42.910 "uuid": "33c46252-4999-5be1-b419-0adb4c459922", 00:35:42.910 "is_configured": true, 00:35:42.910 "data_offset": 256, 00:35:42.910 "data_size": 7936 00:35:42.910 } 00:35:42.910 ] 00:35:42.910 }' 00:35:42.910 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:35:42.910 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:35:42.910 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:35:42.910 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:35:42.910 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:35:42.910 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:42.911 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:42.911 [2024-11-20 13:55:49.636667] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:35:42.911 [2024-11-20 13:55:49.648333] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:35:42.911 [2024-11-20 13:55:49.648381] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:35:42.911 [2024-11-20 13:55:49.648395] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:35:42.911 [2024-11-20 13:55:49.648401] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:35:42.911 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:42.911 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:35:42.911 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:42.911 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:42.911 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:42.911 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:42.911 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:35:42.911 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:42.911 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:42.911 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:42.911 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:42.911 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:42.911 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:42.911 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:42.911 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:42.911 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:42.911 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:42.911 "name": "raid_bdev1", 00:35:42.911 "uuid": "c58e332f-a9bb-468c-8c1d-e91086adebaa", 00:35:42.911 "strip_size_kb": 0, 00:35:42.911 "state": "online", 00:35:42.911 "raid_level": "raid1", 00:35:42.911 "superblock": true, 00:35:42.911 "num_base_bdevs": 2, 00:35:42.911 "num_base_bdevs_discovered": 1, 00:35:42.911 "num_base_bdevs_operational": 1, 00:35:42.911 "base_bdevs_list": [ 00:35:42.911 { 00:35:42.911 "name": null, 00:35:42.911 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:42.911 "is_configured": false, 00:35:42.911 "data_offset": 0, 00:35:42.911 "data_size": 7936 00:35:42.911 }, 00:35:42.911 { 00:35:42.911 "name": "BaseBdev2", 00:35:42.911 "uuid": "33c46252-4999-5be1-b419-0adb4c459922", 00:35:42.911 "is_configured": true, 00:35:42.911 "data_offset": 256, 00:35:42.911 "data_size": 7936 00:35:42.911 } 00:35:42.911 ] 00:35:42.911 }' 00:35:42.911 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:42.911 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:43.169 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:35:43.169 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:35:43.169 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:35:43.169 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:35:43.169 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:35:43.169 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:43.169 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:43.169 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:43.169 13:55:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:43.169 13:55:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:43.169 13:55:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:35:43.169 "name": "raid_bdev1", 00:35:43.169 "uuid": "c58e332f-a9bb-468c-8c1d-e91086adebaa", 00:35:43.169 "strip_size_kb": 0, 00:35:43.169 "state": "online", 00:35:43.169 "raid_level": "raid1", 00:35:43.169 "superblock": true, 00:35:43.169 "num_base_bdevs": 2, 00:35:43.169 "num_base_bdevs_discovered": 1, 00:35:43.169 "num_base_bdevs_operational": 1, 00:35:43.169 "base_bdevs_list": [ 00:35:43.169 { 00:35:43.169 "name": null, 00:35:43.169 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:43.169 "is_configured": false, 00:35:43.169 "data_offset": 0, 00:35:43.169 "data_size": 7936 00:35:43.169 }, 00:35:43.169 { 00:35:43.169 "name": "BaseBdev2", 00:35:43.169 "uuid": "33c46252-4999-5be1-b419-0adb4c459922", 00:35:43.169 "is_configured": true, 00:35:43.169 "data_offset": 256, 00:35:43.169 "data_size": 7936 00:35:43.169 } 00:35:43.169 ] 00:35:43.169 }' 00:35:43.169 13:55:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:35:43.169 13:55:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:35:43.169 13:55:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:35:43.169 13:55:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:35:43.169 13:55:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:35:43.169 13:55:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:43.169 13:55:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:43.169 13:55:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:43.169 13:55:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:35:43.170 13:55:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:43.170 13:55:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:43.170 [2024-11-20 13:55:50.107643] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:35:43.170 [2024-11-20 13:55:50.107696] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:35:43.170 [2024-11-20 13:55:50.107716] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:35:43.170 [2024-11-20 13:55:50.107723] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:35:43.170 [2024-11-20 13:55:50.107894] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:35:43.170 [2024-11-20 13:55:50.107904] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:35:43.170 [2024-11-20 13:55:50.107948] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:35:43.170 [2024-11-20 13:55:50.107961] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:35:43.170 [2024-11-20 13:55:50.107970] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:35:43.170 [2024-11-20 13:55:50.107988] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:35:43.170 BaseBdev1 00:35:43.170 13:55:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:43.170 13:55:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@775 -- # sleep 1 00:35:44.229 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:35:44.229 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:44.229 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:44.229 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:44.229 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:44.229 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:35:44.229 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:44.229 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:44.229 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:44.229 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:44.229 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:44.229 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:44.229 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:44.229 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:44.229 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:44.229 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:44.229 "name": "raid_bdev1", 00:35:44.229 "uuid": "c58e332f-a9bb-468c-8c1d-e91086adebaa", 00:35:44.229 "strip_size_kb": 0, 00:35:44.229 "state": "online", 00:35:44.229 "raid_level": "raid1", 00:35:44.229 "superblock": true, 00:35:44.229 "num_base_bdevs": 2, 00:35:44.229 "num_base_bdevs_discovered": 1, 00:35:44.229 "num_base_bdevs_operational": 1, 00:35:44.229 "base_bdevs_list": [ 00:35:44.229 { 00:35:44.229 "name": null, 00:35:44.229 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:44.229 "is_configured": false, 00:35:44.229 "data_offset": 0, 00:35:44.229 "data_size": 7936 00:35:44.229 }, 00:35:44.229 { 00:35:44.229 "name": "BaseBdev2", 00:35:44.229 "uuid": "33c46252-4999-5be1-b419-0adb4c459922", 00:35:44.229 "is_configured": true, 00:35:44.229 "data_offset": 256, 00:35:44.229 "data_size": 7936 00:35:44.229 } 00:35:44.229 ] 00:35:44.229 }' 00:35:44.229 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:44.229 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:35:44.488 "name": "raid_bdev1", 00:35:44.488 "uuid": "c58e332f-a9bb-468c-8c1d-e91086adebaa", 00:35:44.488 "strip_size_kb": 0, 00:35:44.488 "state": "online", 00:35:44.488 "raid_level": "raid1", 00:35:44.488 "superblock": true, 00:35:44.488 "num_base_bdevs": 2, 00:35:44.488 "num_base_bdevs_discovered": 1, 00:35:44.488 "num_base_bdevs_operational": 1, 00:35:44.488 "base_bdevs_list": [ 00:35:44.488 { 00:35:44.488 "name": null, 00:35:44.488 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:44.488 "is_configured": false, 00:35:44.488 "data_offset": 0, 00:35:44.488 "data_size": 7936 00:35:44.488 }, 00:35:44.488 { 00:35:44.488 "name": "BaseBdev2", 00:35:44.488 "uuid": "33c46252-4999-5be1-b419-0adb4c459922", 00:35:44.488 "is_configured": true, 00:35:44.488 "data_offset": 256, 00:35:44.488 "data_size": 7936 00:35:44.488 } 00:35:44.488 ] 00:35:44.488 }' 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@652 -- # local es=0 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:44.488 [2024-11-20 13:55:51.527945] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:35:44.488 [2024-11-20 13:55:51.528205] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:35:44.488 [2024-11-20 13:55:51.528228] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:35:44.488 request: 00:35:44.488 { 00:35:44.488 "base_bdev": "BaseBdev1", 00:35:44.488 "raid_bdev": "raid_bdev1", 00:35:44.488 "method": "bdev_raid_add_base_bdev", 00:35:44.488 "req_id": 1 00:35:44.488 } 00:35:44.488 Got JSON-RPC error response 00:35:44.488 response: 00:35:44.488 { 00:35:44.488 "code": -22, 00:35:44.488 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:35:44.488 } 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@655 -- # es=1 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:35:44.488 13:55:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@779 -- # sleep 1 00:35:45.862 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:35:45.862 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:35:45.862 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:35:45.862 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:35:45.862 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:35:45.862 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:35:45.862 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:35:45.862 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:35:45.862 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:35:45.862 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:35:45.862 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:45.862 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:45.862 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:45.863 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:45.863 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:45.863 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:35:45.863 "name": "raid_bdev1", 00:35:45.863 "uuid": "c58e332f-a9bb-468c-8c1d-e91086adebaa", 00:35:45.863 "strip_size_kb": 0, 00:35:45.863 "state": "online", 00:35:45.863 "raid_level": "raid1", 00:35:45.863 "superblock": true, 00:35:45.863 "num_base_bdevs": 2, 00:35:45.863 "num_base_bdevs_discovered": 1, 00:35:45.863 "num_base_bdevs_operational": 1, 00:35:45.863 "base_bdevs_list": [ 00:35:45.863 { 00:35:45.863 "name": null, 00:35:45.863 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:45.863 "is_configured": false, 00:35:45.863 "data_offset": 0, 00:35:45.863 "data_size": 7936 00:35:45.863 }, 00:35:45.863 { 00:35:45.863 "name": "BaseBdev2", 00:35:45.863 "uuid": "33c46252-4999-5be1-b419-0adb4c459922", 00:35:45.863 "is_configured": true, 00:35:45.863 "data_offset": 256, 00:35:45.863 "data_size": 7936 00:35:45.863 } 00:35:45.863 ] 00:35:45.863 }' 00:35:45.863 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:35:45.863 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:45.863 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:35:45.863 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:35:45.863 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:35:45.863 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:35:45.863 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:35:45.863 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:35:45.863 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:35:45.863 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:45.863 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:45.863 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:45.863 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:35:45.863 "name": "raid_bdev1", 00:35:45.863 "uuid": "c58e332f-a9bb-468c-8c1d-e91086adebaa", 00:35:45.863 "strip_size_kb": 0, 00:35:45.863 "state": "online", 00:35:45.863 "raid_level": "raid1", 00:35:45.863 "superblock": true, 00:35:45.863 "num_base_bdevs": 2, 00:35:45.863 "num_base_bdevs_discovered": 1, 00:35:45.863 "num_base_bdevs_operational": 1, 00:35:45.863 "base_bdevs_list": [ 00:35:45.863 { 00:35:45.863 "name": null, 00:35:45.863 "uuid": "00000000-0000-0000-0000-000000000000", 00:35:45.863 "is_configured": false, 00:35:45.863 "data_offset": 0, 00:35:45.863 "data_size": 7936 00:35:45.863 }, 00:35:45.863 { 00:35:45.863 "name": "BaseBdev2", 00:35:45.863 "uuid": "33c46252-4999-5be1-b419-0adb4c459922", 00:35:45.863 "is_configured": true, 00:35:45.863 "data_offset": 256, 00:35:45.863 "data_size": 7936 00:35:45.863 } 00:35:45.863 ] 00:35:45.863 }' 00:35:45.863 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:35:45.863 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:35:45.863 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:35:46.120 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:35:46.120 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@784 -- # killprocess 86378 00:35:46.121 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@954 -- # '[' -z 86378 ']' 00:35:46.121 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@958 -- # kill -0 86378 00:35:46.121 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # uname 00:35:46.121 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:35:46.121 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 86378 00:35:46.121 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:35:46.121 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:35:46.121 killing process with pid 86378 00:35:46.121 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@972 -- # echo 'killing process with pid 86378' 00:35:46.121 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@973 -- # kill 86378 00:35:46.121 Received shutdown signal, test time was about 60.000000 seconds 00:35:46.121 00:35:46.121 Latency(us) 00:35:46.121 [2024-11-20T13:55:53.180Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:35:46.121 [2024-11-20T13:55:53.180Z] =================================================================================================================== 00:35:46.121 [2024-11-20T13:55:53.180Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:35:46.121 [2024-11-20 13:55:52.960054] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:35:46.121 13:55:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@978 -- # wait 86378 00:35:46.121 [2024-11-20 13:55:52.960179] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:35:46.121 [2024-11-20 13:55:52.960228] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:35:46.121 [2024-11-20 13:55:52.960238] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:35:46.121 [2024-11-20 13:55:53.110927] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:35:46.686 13:55:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@786 -- # return 0 00:35:46.686 00:35:46.686 real 0m14.777s 00:35:46.686 user 0m18.718s 00:35:46.686 sys 0m1.150s 00:35:46.686 13:55:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@1130 -- # xtrace_disable 00:35:46.686 ************************************ 00:35:46.686 END TEST raid_rebuild_test_sb_md_interleaved 00:35:46.686 ************************************ 00:35:46.686 13:55:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:35:46.686 13:55:53 bdev_raid -- bdev/bdev_raid.sh@1015 -- # trap - EXIT 00:35:46.686 13:55:53 bdev_raid -- bdev/bdev_raid.sh@1016 -- # cleanup 00:35:46.686 13:55:53 bdev_raid -- bdev/bdev_raid.sh@56 -- # '[' -n 86378 ']' 00:35:46.686 13:55:53 bdev_raid -- bdev/bdev_raid.sh@56 -- # ps -p 86378 00:35:46.944 13:55:53 bdev_raid -- bdev/bdev_raid.sh@60 -- # rm -rf /raidtest 00:35:46.944 ************************************ 00:35:46.944 END TEST bdev_raid 00:35:46.944 ************************************ 00:35:46.944 00:35:46.944 real 9m30.208s 00:35:46.944 user 12m36.257s 00:35:46.944 sys 1m23.469s 00:35:46.944 13:55:53 bdev_raid -- common/autotest_common.sh@1130 -- # xtrace_disable 00:35:46.944 13:55:53 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:35:46.944 13:55:53 -- spdk/autotest.sh@190 -- # run_test spdkcli_raid /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:35:46.944 13:55:53 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:35:46.944 13:55:53 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:35:46.944 13:55:53 -- common/autotest_common.sh@10 -- # set +x 00:35:46.944 ************************************ 00:35:46.944 START TEST spdkcli_raid 00:35:46.944 ************************************ 00:35:46.944 13:55:53 spdkcli_raid -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:35:46.944 * Looking for test storage... 00:35:46.944 * Found test storage at /home/vagrant/spdk_repo/spdk/test/spdkcli 00:35:46.944 13:55:53 spdkcli_raid -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:35:46.944 13:55:53 spdkcli_raid -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:35:46.944 13:55:53 spdkcli_raid -- common/autotest_common.sh@1693 -- # lcov --version 00:35:46.944 13:55:53 spdkcli_raid -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@336 -- # IFS=.-: 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@336 -- # read -ra ver1 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@337 -- # IFS=.-: 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@337 -- # read -ra ver2 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@338 -- # local 'op=<' 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@340 -- # ver1_l=2 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@341 -- # ver2_l=1 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@344 -- # case "$op" in 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@345 -- # : 1 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@364 -- # (( v = 0 )) 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@365 -- # decimal 1 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@353 -- # local d=1 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@355 -- # echo 1 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@365 -- # ver1[v]=1 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@366 -- # decimal 2 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@353 -- # local d=2 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@355 -- # echo 2 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@366 -- # ver2[v]=2 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:35:46.944 13:55:53 spdkcli_raid -- scripts/common.sh@368 -- # return 0 00:35:46.944 13:55:53 spdkcli_raid -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:35:46.944 13:55:53 spdkcli_raid -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:35:46.944 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:46.944 --rc genhtml_branch_coverage=1 00:35:46.944 --rc genhtml_function_coverage=1 00:35:46.944 --rc genhtml_legend=1 00:35:46.944 --rc geninfo_all_blocks=1 00:35:46.944 --rc geninfo_unexecuted_blocks=1 00:35:46.944 00:35:46.944 ' 00:35:46.944 13:55:53 spdkcli_raid -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:35:46.944 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:46.944 --rc genhtml_branch_coverage=1 00:35:46.944 --rc genhtml_function_coverage=1 00:35:46.944 --rc genhtml_legend=1 00:35:46.944 --rc geninfo_all_blocks=1 00:35:46.944 --rc geninfo_unexecuted_blocks=1 00:35:46.944 00:35:46.944 ' 00:35:46.944 13:55:53 spdkcli_raid -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:35:46.944 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:46.944 --rc genhtml_branch_coverage=1 00:35:46.944 --rc genhtml_function_coverage=1 00:35:46.944 --rc genhtml_legend=1 00:35:46.944 --rc geninfo_all_blocks=1 00:35:46.944 --rc geninfo_unexecuted_blocks=1 00:35:46.944 00:35:46.944 ' 00:35:46.944 13:55:53 spdkcli_raid -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:35:46.944 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:46.944 --rc genhtml_branch_coverage=1 00:35:46.944 --rc genhtml_function_coverage=1 00:35:46.944 --rc genhtml_legend=1 00:35:46.944 --rc geninfo_all_blocks=1 00:35:46.944 --rc geninfo_unexecuted_blocks=1 00:35:46.944 00:35:46.944 ' 00:35:46.944 13:55:53 spdkcli_raid -- spdkcli/raid.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:35:46.944 13:55:53 spdkcli_raid -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:35:46.945 13:55:53 spdkcli_raid -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:35:46.945 13:55:53 spdkcli_raid -- spdkcli/raid.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/iscsi_tgt/common.sh 00:35:46.945 13:55:53 spdkcli_raid -- iscsi_tgt/common.sh@9 -- # ISCSI_BRIDGE=iscsi_br 00:35:46.945 13:55:53 spdkcli_raid -- iscsi_tgt/common.sh@10 -- # INITIATOR_INTERFACE=spdk_init_int 00:35:46.945 13:55:53 spdkcli_raid -- iscsi_tgt/common.sh@11 -- # INITIATOR_BRIDGE=init_br 00:35:46.945 13:55:53 spdkcli_raid -- iscsi_tgt/common.sh@12 -- # TARGET_NAMESPACE=spdk_iscsi_ns 00:35:46.945 13:55:53 spdkcli_raid -- iscsi_tgt/common.sh@13 -- # TARGET_NS_CMD=(ip netns exec "$TARGET_NAMESPACE") 00:35:46.945 13:55:53 spdkcli_raid -- iscsi_tgt/common.sh@14 -- # TARGET_INTERFACE=spdk_tgt_int 00:35:46.945 13:55:53 spdkcli_raid -- iscsi_tgt/common.sh@15 -- # TARGET_INTERFACE2=spdk_tgt_int2 00:35:46.945 13:55:53 spdkcli_raid -- iscsi_tgt/common.sh@16 -- # TARGET_BRIDGE=tgt_br 00:35:46.945 13:55:53 spdkcli_raid -- iscsi_tgt/common.sh@17 -- # TARGET_BRIDGE2=tgt_br2 00:35:46.945 13:55:53 spdkcli_raid -- iscsi_tgt/common.sh@20 -- # TARGET_IP=10.0.0.1 00:35:46.945 13:55:53 spdkcli_raid -- iscsi_tgt/common.sh@21 -- # TARGET_IP2=10.0.0.3 00:35:46.945 13:55:53 spdkcli_raid -- iscsi_tgt/common.sh@22 -- # INITIATOR_IP=10.0.0.2 00:35:46.945 13:55:53 spdkcli_raid -- iscsi_tgt/common.sh@23 -- # ISCSI_PORT=3260 00:35:46.945 13:55:53 spdkcli_raid -- iscsi_tgt/common.sh@24 -- # NETMASK=10.0.0.2/32 00:35:46.945 13:55:53 spdkcli_raid -- iscsi_tgt/common.sh@25 -- # INITIATOR_TAG=2 00:35:46.945 13:55:53 spdkcli_raid -- iscsi_tgt/common.sh@26 -- # INITIATOR_NAME=ANY 00:35:46.945 13:55:53 spdkcli_raid -- iscsi_tgt/common.sh@27 -- # PORTAL_TAG=1 00:35:46.945 13:55:53 spdkcli_raid -- iscsi_tgt/common.sh@28 -- # ISCSI_APP=("${TARGET_NS_CMD[@]}" "${ISCSI_APP[@]}") 00:35:46.945 13:55:53 spdkcli_raid -- iscsi_tgt/common.sh@29 -- # ISCSI_TEST_CORE_MASK=0xF 00:35:46.945 13:55:53 spdkcli_raid -- spdkcli/raid.sh@12 -- # MATCH_FILE=spdkcli_raid.test 00:35:46.945 13:55:53 spdkcli_raid -- spdkcli/raid.sh@13 -- # SPDKCLI_BRANCH=/bdevs 00:35:46.945 13:55:53 spdkcli_raid -- spdkcli/raid.sh@14 -- # dirname /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:35:46.945 13:55:53 spdkcli_raid -- spdkcli/raid.sh@14 -- # readlink -f /home/vagrant/spdk_repo/spdk/test/spdkcli 00:35:46.945 13:55:53 spdkcli_raid -- spdkcli/raid.sh@14 -- # testdir=/home/vagrant/spdk_repo/spdk/test/spdkcli 00:35:46.945 13:55:53 spdkcli_raid -- spdkcli/raid.sh@15 -- # . /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:35:46.945 13:55:53 spdkcli_raid -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:35:46.945 13:55:53 spdkcli_raid -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:35:46.945 13:55:53 spdkcli_raid -- spdkcli/raid.sh@17 -- # trap cleanup EXIT 00:35:46.945 13:55:53 spdkcli_raid -- spdkcli/raid.sh@19 -- # timing_enter run_spdk_tgt 00:35:46.945 13:55:53 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:35:46.945 13:55:53 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:35:46.945 13:55:53 spdkcli_raid -- spdkcli/raid.sh@20 -- # run_spdk_tgt 00:35:46.945 13:55:53 spdkcli_raid -- spdkcli/common.sh@27 -- # spdk_tgt_pid=87032 00:35:46.945 13:55:53 spdkcli_raid -- spdkcli/common.sh@28 -- # waitforlisten 87032 00:35:46.945 13:55:53 spdkcli_raid -- spdkcli/common.sh@26 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:35:46.945 13:55:53 spdkcli_raid -- common/autotest_common.sh@835 -- # '[' -z 87032 ']' 00:35:46.945 13:55:53 spdkcli_raid -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:35:46.945 13:55:53 spdkcli_raid -- common/autotest_common.sh@840 -- # local max_retries=100 00:35:46.945 13:55:53 spdkcli_raid -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:35:46.945 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:35:46.945 13:55:53 spdkcli_raid -- common/autotest_common.sh@844 -- # xtrace_disable 00:35:46.945 13:55:53 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:35:47.203 [2024-11-20 13:55:54.024628] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:35:47.203 [2024-11-20 13:55:54.024884] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87032 ] 00:35:47.203 [2024-11-20 13:55:54.187864] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:35:47.460 [2024-11-20 13:55:54.300697] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:35:47.460 [2024-11-20 13:55:54.300831] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:35:48.025 13:55:54 spdkcli_raid -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:35:48.025 13:55:54 spdkcli_raid -- common/autotest_common.sh@868 -- # return 0 00:35:48.025 13:55:54 spdkcli_raid -- spdkcli/raid.sh@21 -- # timing_exit run_spdk_tgt 00:35:48.025 13:55:54 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:35:48.025 13:55:54 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:35:48.025 13:55:54 spdkcli_raid -- spdkcli/raid.sh@23 -- # timing_enter spdkcli_create_malloc 00:35:48.025 13:55:54 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:35:48.025 13:55:54 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:35:48.025 13:55:54 spdkcli_raid -- spdkcli/raid.sh@26 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc create 8 512 Malloc1'\'' '\''Malloc1'\'' True 00:35:48.025 '\''/bdevs/malloc create 8 512 Malloc2'\'' '\''Malloc2'\'' True 00:35:48.025 ' 00:35:49.399 Executing command: ['/bdevs/malloc create 8 512 Malloc1', 'Malloc1', True] 00:35:49.399 Executing command: ['/bdevs/malloc create 8 512 Malloc2', 'Malloc2', True] 00:35:49.657 13:55:56 spdkcli_raid -- spdkcli/raid.sh@27 -- # timing_exit spdkcli_create_malloc 00:35:49.657 13:55:56 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:35:49.657 13:55:56 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:35:49.657 13:55:56 spdkcli_raid -- spdkcli/raid.sh@29 -- # timing_enter spdkcli_create_raid 00:35:49.657 13:55:56 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:35:49.657 13:55:56 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:35:49.657 13:55:56 spdkcli_raid -- spdkcli/raid.sh@31 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/raid_volume create testraid 0 "Malloc1 Malloc2" 4'\'' '\''testraid'\'' True 00:35:49.657 ' 00:35:50.591 Executing command: ['/bdevs/raid_volume create testraid 0 "Malloc1 Malloc2" 4', 'testraid', True] 00:35:50.591 13:55:57 spdkcli_raid -- spdkcli/raid.sh@32 -- # timing_exit spdkcli_create_raid 00:35:50.591 13:55:57 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:35:50.591 13:55:57 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:35:50.849 13:55:57 spdkcli_raid -- spdkcli/raid.sh@34 -- # timing_enter spdkcli_check_match 00:35:50.849 13:55:57 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:35:50.849 13:55:57 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:35:50.849 13:55:57 spdkcli_raid -- spdkcli/raid.sh@35 -- # check_match 00:35:50.850 13:55:57 spdkcli_raid -- spdkcli/common.sh@44 -- # /home/vagrant/spdk_repo/spdk/scripts/spdkcli.py ll /bdevs 00:35:51.108 13:55:58 spdkcli_raid -- spdkcli/common.sh@45 -- # /home/vagrant/spdk_repo/spdk/test/app/match/match /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_raid.test.match 00:35:51.366 13:55:58 spdkcli_raid -- spdkcli/common.sh@46 -- # rm -f /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_raid.test 00:35:51.366 13:55:58 spdkcli_raid -- spdkcli/raid.sh@36 -- # timing_exit spdkcli_check_match 00:35:51.366 13:55:58 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:35:51.366 13:55:58 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:35:51.366 13:55:58 spdkcli_raid -- spdkcli/raid.sh@38 -- # timing_enter spdkcli_delete_raid 00:35:51.366 13:55:58 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:35:51.366 13:55:58 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:35:51.366 13:55:58 spdkcli_raid -- spdkcli/raid.sh@40 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/raid_volume delete testraid'\'' '\'''\'' True 00:35:51.366 ' 00:35:52.385 Executing command: ['/bdevs/raid_volume delete testraid', '', True] 00:35:52.385 13:55:59 spdkcli_raid -- spdkcli/raid.sh@41 -- # timing_exit spdkcli_delete_raid 00:35:52.385 13:55:59 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:35:52.385 13:55:59 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:35:52.385 13:55:59 spdkcli_raid -- spdkcli/raid.sh@43 -- # timing_enter spdkcli_delete_malloc 00:35:52.385 13:55:59 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:35:52.385 13:55:59 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:35:52.385 13:55:59 spdkcli_raid -- spdkcli/raid.sh@46 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc delete Malloc1'\'' '\'''\'' True 00:35:52.385 '\''/bdevs/malloc delete Malloc2'\'' '\'''\'' True 00:35:52.385 ' 00:35:53.757 Executing command: ['/bdevs/malloc delete Malloc1', '', True] 00:35:53.757 Executing command: ['/bdevs/malloc delete Malloc2', '', True] 00:35:53.757 13:56:00 spdkcli_raid -- spdkcli/raid.sh@47 -- # timing_exit spdkcli_delete_malloc 00:35:53.757 13:56:00 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:35:53.757 13:56:00 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:35:53.757 13:56:00 spdkcli_raid -- spdkcli/raid.sh@49 -- # killprocess 87032 00:35:53.757 13:56:00 spdkcli_raid -- common/autotest_common.sh@954 -- # '[' -z 87032 ']' 00:35:53.757 13:56:00 spdkcli_raid -- common/autotest_common.sh@958 -- # kill -0 87032 00:35:53.757 13:56:00 spdkcli_raid -- common/autotest_common.sh@959 -- # uname 00:35:53.757 13:56:00 spdkcli_raid -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:35:53.757 13:56:00 spdkcli_raid -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 87032 00:35:53.757 killing process with pid 87032 00:35:53.757 13:56:00 spdkcli_raid -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:35:53.757 13:56:00 spdkcli_raid -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:35:53.757 13:56:00 spdkcli_raid -- common/autotest_common.sh@972 -- # echo 'killing process with pid 87032' 00:35:53.757 13:56:00 spdkcli_raid -- common/autotest_common.sh@973 -- # kill 87032 00:35:53.757 13:56:00 spdkcli_raid -- common/autotest_common.sh@978 -- # wait 87032 00:35:55.130 13:56:02 spdkcli_raid -- spdkcli/raid.sh@1 -- # cleanup 00:35:55.130 13:56:02 spdkcli_raid -- spdkcli/common.sh@10 -- # '[' -n 87032 ']' 00:35:55.130 13:56:02 spdkcli_raid -- spdkcli/common.sh@11 -- # killprocess 87032 00:35:55.130 13:56:02 spdkcli_raid -- common/autotest_common.sh@954 -- # '[' -z 87032 ']' 00:35:55.130 Process with pid 87032 is not found 00:35:55.130 13:56:02 spdkcli_raid -- common/autotest_common.sh@958 -- # kill -0 87032 00:35:55.130 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 958: kill: (87032) - No such process 00:35:55.130 13:56:02 spdkcli_raid -- common/autotest_common.sh@981 -- # echo 'Process with pid 87032 is not found' 00:35:55.130 13:56:02 spdkcli_raid -- spdkcli/common.sh@13 -- # '[' -n '' ']' 00:35:55.130 13:56:02 spdkcli_raid -- spdkcli/common.sh@16 -- # '[' -n '' ']' 00:35:55.130 13:56:02 spdkcli_raid -- spdkcli/common.sh@19 -- # '[' -n '' ']' 00:35:55.130 13:56:02 spdkcli_raid -- spdkcli/common.sh@22 -- # rm -f /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_raid.test /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_details_vhost.test /tmp/sample_aio 00:35:55.130 ************************************ 00:35:55.130 END TEST spdkcli_raid 00:35:55.130 ************************************ 00:35:55.130 00:35:55.130 real 0m8.246s 00:35:55.130 user 0m17.040s 00:35:55.130 sys 0m0.831s 00:35:55.130 13:56:02 spdkcli_raid -- common/autotest_common.sh@1130 -- # xtrace_disable 00:35:55.130 13:56:02 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:35:55.130 13:56:02 -- spdk/autotest.sh@191 -- # run_test blockdev_raid5f /home/vagrant/spdk_repo/spdk/test/bdev/blockdev.sh raid5f 00:35:55.130 13:56:02 -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:35:55.130 13:56:02 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:35:55.130 13:56:02 -- common/autotest_common.sh@10 -- # set +x 00:35:55.130 ************************************ 00:35:55.130 START TEST blockdev_raid5f 00:35:55.130 ************************************ 00:35:55.130 13:56:02 blockdev_raid5f -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/bdev/blockdev.sh raid5f 00:35:55.130 * Looking for test storage... 00:35:55.130 * Found test storage at /home/vagrant/spdk_repo/spdk/test/bdev 00:35:55.130 13:56:02 blockdev_raid5f -- common/autotest_common.sh@1692 -- # [[ y == y ]] 00:35:55.130 13:56:02 blockdev_raid5f -- common/autotest_common.sh@1693 -- # awk '{print $NF}' 00:35:55.130 13:56:02 blockdev_raid5f -- common/autotest_common.sh@1693 -- # lcov --version 00:35:55.389 13:56:02 blockdev_raid5f -- common/autotest_common.sh@1693 -- # lt 1.15 2 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@333 -- # local ver1 ver1_l 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@334 -- # local ver2 ver2_l 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@336 -- # IFS=.-: 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@336 -- # read -ra ver1 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@337 -- # IFS=.-: 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@337 -- # read -ra ver2 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@338 -- # local 'op=<' 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@340 -- # ver1_l=2 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@341 -- # ver2_l=1 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@344 -- # case "$op" in 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@345 -- # : 1 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@364 -- # (( v = 0 )) 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@365 -- # decimal 1 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@353 -- # local d=1 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@355 -- # echo 1 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@365 -- # ver1[v]=1 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@366 -- # decimal 2 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@353 -- # local d=2 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@355 -- # echo 2 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@366 -- # ver2[v]=2 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:35:55.389 13:56:02 blockdev_raid5f -- scripts/common.sh@368 -- # return 0 00:35:55.389 13:56:02 blockdev_raid5f -- common/autotest_common.sh@1694 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:35:55.389 13:56:02 blockdev_raid5f -- common/autotest_common.sh@1706 -- # export 'LCOV_OPTS= 00:35:55.389 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:55.389 --rc genhtml_branch_coverage=1 00:35:55.389 --rc genhtml_function_coverage=1 00:35:55.389 --rc genhtml_legend=1 00:35:55.389 --rc geninfo_all_blocks=1 00:35:55.389 --rc geninfo_unexecuted_blocks=1 00:35:55.389 00:35:55.389 ' 00:35:55.389 13:56:02 blockdev_raid5f -- common/autotest_common.sh@1706 -- # LCOV_OPTS=' 00:35:55.389 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:55.389 --rc genhtml_branch_coverage=1 00:35:55.389 --rc genhtml_function_coverage=1 00:35:55.389 --rc genhtml_legend=1 00:35:55.389 --rc geninfo_all_blocks=1 00:35:55.389 --rc geninfo_unexecuted_blocks=1 00:35:55.389 00:35:55.389 ' 00:35:55.389 13:56:02 blockdev_raid5f -- common/autotest_common.sh@1707 -- # export 'LCOV=lcov 00:35:55.389 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:55.389 --rc genhtml_branch_coverage=1 00:35:55.389 --rc genhtml_function_coverage=1 00:35:55.389 --rc genhtml_legend=1 00:35:55.389 --rc geninfo_all_blocks=1 00:35:55.389 --rc geninfo_unexecuted_blocks=1 00:35:55.389 00:35:55.389 ' 00:35:55.389 13:56:02 blockdev_raid5f -- common/autotest_common.sh@1707 -- # LCOV='lcov 00:35:55.389 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:55.389 --rc genhtml_branch_coverage=1 00:35:55.389 --rc genhtml_function_coverage=1 00:35:55.389 --rc genhtml_legend=1 00:35:55.389 --rc geninfo_all_blocks=1 00:35:55.389 --rc geninfo_unexecuted_blocks=1 00:35:55.389 00:35:55.389 ' 00:35:55.389 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:35:55.389 13:56:02 blockdev_raid5f -- bdev/nbd_common.sh@6 -- # set -e 00:35:55.389 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@12 -- # rpc_py=rpc_cmd 00:35:55.389 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@13 -- # conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:35:55.389 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@14 -- # nonenclosed_conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json 00:35:55.389 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@15 -- # nonarray_conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json 00:35:55.389 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@17 -- # export RPC_PIPE_TIMEOUT=30 00:35:55.389 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@17 -- # RPC_PIPE_TIMEOUT=30 00:35:55.389 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@20 -- # : 00:35:55.389 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@707 -- # QOS_DEV_1=Malloc_0 00:35:55.389 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@708 -- # QOS_DEV_2=Null_1 00:35:55.389 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@709 -- # QOS_RUN_TIME=5 00:35:55.389 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@711 -- # uname -s 00:35:55.389 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:35:55.389 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@711 -- # '[' Linux = Linux ']' 00:35:55.389 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@713 -- # PRE_RESERVED_MEM=0 00:35:55.389 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@719 -- # test_type=raid5f 00:35:55.389 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@720 -- # crypto_device= 00:35:55.389 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@721 -- # dek= 00:35:55.389 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@722 -- # env_ctx= 00:35:55.389 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@723 -- # wait_for_rpc= 00:35:55.389 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@724 -- # '[' -n '' ']' 00:35:55.390 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@727 -- # [[ raid5f == bdev ]] 00:35:55.390 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@727 -- # [[ raid5f == crypto_* ]] 00:35:55.390 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@730 -- # start_spdk_tgt 00:35:55.390 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@47 -- # spdk_tgt_pid=87291 00:35:55.390 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@48 -- # trap 'killprocess "$spdk_tgt_pid"; exit 1' SIGINT SIGTERM EXIT 00:35:55.390 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@49 -- # waitforlisten 87291 00:35:55.390 13:56:02 blockdev_raid5f -- common/autotest_common.sh@835 -- # '[' -z 87291 ']' 00:35:55.390 13:56:02 blockdev_raid5f -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:35:55.390 13:56:02 blockdev_raid5f -- common/autotest_common.sh@840 -- # local max_retries=100 00:35:55.390 13:56:02 blockdev_raid5f -- bdev/blockdev.sh@46 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt '' '' 00:35:55.390 13:56:02 blockdev_raid5f -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:35:55.390 13:56:02 blockdev_raid5f -- common/autotest_common.sh@844 -- # xtrace_disable 00:35:55.390 13:56:02 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:35:55.390 [2024-11-20 13:56:02.308797] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:35:55.390 [2024-11-20 13:56:02.308932] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87291 ] 00:35:55.648 [2024-11-20 13:56:02.466746] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:35:55.648 [2024-11-20 13:56:02.566312] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:35:56.213 13:56:03 blockdev_raid5f -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:35:56.213 13:56:03 blockdev_raid5f -- common/autotest_common.sh@868 -- # return 0 00:35:56.213 13:56:03 blockdev_raid5f -- bdev/blockdev.sh@731 -- # case "$test_type" in 00:35:56.213 13:56:03 blockdev_raid5f -- bdev/blockdev.sh@763 -- # setup_raid5f_conf 00:35:56.214 13:56:03 blockdev_raid5f -- bdev/blockdev.sh@279 -- # rpc_cmd 00:35:56.214 13:56:03 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:56.214 13:56:03 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:35:56.214 Malloc0 00:35:56.214 Malloc1 00:35:56.214 Malloc2 00:35:56.214 13:56:03 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:56.214 13:56:03 blockdev_raid5f -- bdev/blockdev.sh@774 -- # rpc_cmd bdev_wait_for_examine 00:35:56.214 13:56:03 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:56.214 13:56:03 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:35:56.214 13:56:03 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:56.214 13:56:03 blockdev_raid5f -- bdev/blockdev.sh@777 -- # cat 00:35:56.214 13:56:03 blockdev_raid5f -- bdev/blockdev.sh@777 -- # rpc_cmd save_subsystem_config -n accel 00:35:56.214 13:56:03 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:56.214 13:56:03 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:35:56.214 13:56:03 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:56.214 13:56:03 blockdev_raid5f -- bdev/blockdev.sh@777 -- # rpc_cmd save_subsystem_config -n bdev 00:35:56.214 13:56:03 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:56.214 13:56:03 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:35:56.214 13:56:03 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:56.472 13:56:03 blockdev_raid5f -- bdev/blockdev.sh@777 -- # rpc_cmd save_subsystem_config -n iobuf 00:35:56.472 13:56:03 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:56.472 13:56:03 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:35:56.472 13:56:03 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:56.472 13:56:03 blockdev_raid5f -- bdev/blockdev.sh@785 -- # mapfile -t bdevs 00:35:56.472 13:56:03 blockdev_raid5f -- bdev/blockdev.sh@785 -- # jq -r '.[] | select(.claimed == false)' 00:35:56.472 13:56:03 blockdev_raid5f -- bdev/blockdev.sh@785 -- # rpc_cmd bdev_get_bdevs 00:35:56.472 13:56:03 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:56.472 13:56:03 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:35:56.472 13:56:03 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:56.472 13:56:03 blockdev_raid5f -- bdev/blockdev.sh@786 -- # mapfile -t bdevs_name 00:35:56.472 13:56:03 blockdev_raid5f -- bdev/blockdev.sh@786 -- # printf '%s\n' '{' ' "name": "raid5f",' ' "aliases": [' ' "1143f858-e763-4400-8e6a-dd7ee4b79357"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "1143f858-e763-4400-8e6a-dd7ee4b79357",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "raid": {' ' "uuid": "1143f858-e763-4400-8e6a-dd7ee4b79357",' ' "strip_size_kb": 2,' ' "state": "online",' ' "raid_level": "raid5f",' ' "superblock": false,' ' "num_base_bdevs": 3,' ' "num_base_bdevs_discovered": 3,' ' "num_base_bdevs_operational": 3,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc0",' ' "uuid": "2be9841d-5c24-445e-9e69-8fd81a7b244f",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc1",' ' "uuid": "0ea55a6d-7d4c-47c3-97ad-326b81a47106",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc2",' ' "uuid": "1e0b630f-0da7-4ff0-a129-ea28b016558c",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' 00:35:56.472 13:56:03 blockdev_raid5f -- bdev/blockdev.sh@786 -- # jq -r .name 00:35:56.472 13:56:03 blockdev_raid5f -- bdev/blockdev.sh@787 -- # bdev_list=("${bdevs_name[@]}") 00:35:56.472 13:56:03 blockdev_raid5f -- bdev/blockdev.sh@789 -- # hello_world_bdev=raid5f 00:35:56.472 13:56:03 blockdev_raid5f -- bdev/blockdev.sh@790 -- # trap - SIGINT SIGTERM EXIT 00:35:56.472 13:56:03 blockdev_raid5f -- bdev/blockdev.sh@791 -- # killprocess 87291 00:35:56.472 13:56:03 blockdev_raid5f -- common/autotest_common.sh@954 -- # '[' -z 87291 ']' 00:35:56.472 13:56:03 blockdev_raid5f -- common/autotest_common.sh@958 -- # kill -0 87291 00:35:56.472 13:56:03 blockdev_raid5f -- common/autotest_common.sh@959 -- # uname 00:35:56.472 13:56:03 blockdev_raid5f -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:35:56.472 13:56:03 blockdev_raid5f -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 87291 00:35:56.472 killing process with pid 87291 00:35:56.472 13:56:03 blockdev_raid5f -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:35:56.472 13:56:03 blockdev_raid5f -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:35:56.472 13:56:03 blockdev_raid5f -- common/autotest_common.sh@972 -- # echo 'killing process with pid 87291' 00:35:56.472 13:56:03 blockdev_raid5f -- common/autotest_common.sh@973 -- # kill 87291 00:35:56.472 13:56:03 blockdev_raid5f -- common/autotest_common.sh@978 -- # wait 87291 00:35:57.850 13:56:04 blockdev_raid5f -- bdev/blockdev.sh@795 -- # trap cleanup SIGINT SIGTERM EXIT 00:35:57.850 13:56:04 blockdev_raid5f -- bdev/blockdev.sh@797 -- # run_test bdev_hello_world /home/vagrant/spdk_repo/spdk/build/examples/hello_bdev --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -b raid5f '' 00:35:57.850 13:56:04 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:35:57.850 13:56:04 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:35:57.850 13:56:04 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:35:57.850 ************************************ 00:35:57.850 START TEST bdev_hello_world 00:35:57.850 ************************************ 00:35:57.850 13:56:04 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/hello_bdev --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -b raid5f '' 00:35:57.850 [2024-11-20 13:56:04.851434] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:35:57.850 [2024-11-20 13:56:04.851713] [ DPDK EAL parameters: hello_bdev --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87336 ] 00:35:58.110 [2024-11-20 13:56:05.007802] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:35:58.110 [2024-11-20 13:56:05.100888] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:35:58.678 [2024-11-20 13:56:05.456302] hello_bdev.c: 222:hello_start: *NOTICE*: Successfully started the application 00:35:58.678 [2024-11-20 13:56:05.456511] hello_bdev.c: 231:hello_start: *NOTICE*: Opening the bdev raid5f 00:35:58.678 [2024-11-20 13:56:05.456544] hello_bdev.c: 244:hello_start: *NOTICE*: Opening io channel 00:35:58.678 [2024-11-20 13:56:05.457252] hello_bdev.c: 138:hello_write: *NOTICE*: Writing to the bdev 00:35:58.678 [2024-11-20 13:56:05.457376] hello_bdev.c: 117:write_complete: *NOTICE*: bdev io write completed successfully 00:35:58.678 [2024-11-20 13:56:05.457390] hello_bdev.c: 84:hello_read: *NOTICE*: Reading io 00:35:58.678 [2024-11-20 13:56:05.457435] hello_bdev.c: 65:read_complete: *NOTICE*: Read string from bdev : Hello World! 00:35:58.678 00:35:58.678 [2024-11-20 13:56:05.457450] hello_bdev.c: 74:read_complete: *NOTICE*: Stopping app 00:35:59.245 00:35:59.245 real 0m1.410s 00:35:59.245 user 0m1.092s 00:35:59.245 sys 0m0.201s 00:35:59.245 13:56:06 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@1130 -- # xtrace_disable 00:35:59.245 13:56:06 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@10 -- # set +x 00:35:59.245 ************************************ 00:35:59.245 END TEST bdev_hello_world 00:35:59.245 ************************************ 00:35:59.245 13:56:06 blockdev_raid5f -- bdev/blockdev.sh@798 -- # run_test bdev_bounds bdev_bounds '' 00:35:59.245 13:56:06 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:35:59.245 13:56:06 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:35:59.245 13:56:06 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:35:59.245 ************************************ 00:35:59.245 START TEST bdev_bounds 00:35:59.245 ************************************ 00:35:59.245 13:56:06 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@1129 -- # bdev_bounds '' 00:35:59.246 13:56:06 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@289 -- # bdevio_pid=87378 00:35:59.246 13:56:06 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@290 -- # trap 'cleanup; killprocess $bdevio_pid; exit 1' SIGINT SIGTERM EXIT 00:35:59.246 13:56:06 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@291 -- # echo 'Process bdevio pid: 87378' 00:35:59.246 Process bdevio pid: 87378 00:35:59.246 13:56:06 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@288 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdevio/bdevio -w -s 0 --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json '' 00:35:59.246 13:56:06 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@292 -- # waitforlisten 87378 00:35:59.246 13:56:06 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@835 -- # '[' -z 87378 ']' 00:35:59.246 13:56:06 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:35:59.246 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:35:59.246 13:56:06 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@840 -- # local max_retries=100 00:35:59.246 13:56:06 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:35:59.246 13:56:06 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@844 -- # xtrace_disable 00:35:59.246 13:56:06 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:35:59.504 [2024-11-20 13:56:06.305568] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:35:59.504 [2024-11-20 13:56:06.305706] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 -m 0 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87378 ] 00:35:59.504 [2024-11-20 13:56:06.461435] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:35:59.504 [2024-11-20 13:56:06.557124] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:35:59.504 [2024-11-20 13:56:06.557375] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:35:59.504 [2024-11-20 13:56:06.557440] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:36:00.440 13:56:07 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:36:00.440 13:56:07 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@868 -- # return 0 00:36:00.440 13:56:07 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@293 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdevio/tests.py perform_tests 00:36:00.440 I/O targets: 00:36:00.440 raid5f: 131072 blocks of 512 bytes (64 MiB) 00:36:00.440 00:36:00.440 00:36:00.440 CUnit - A unit testing framework for C - Version 2.1-3 00:36:00.440 http://cunit.sourceforge.net/ 00:36:00.440 00:36:00.440 00:36:00.440 Suite: bdevio tests on: raid5f 00:36:00.440 Test: blockdev write read block ...passed 00:36:00.440 Test: blockdev write zeroes read block ...passed 00:36:00.440 Test: blockdev write zeroes read no split ...passed 00:36:00.440 Test: blockdev write zeroes read split ...passed 00:36:00.440 Test: blockdev write zeroes read split partial ...passed 00:36:00.440 Test: blockdev reset ...passed 00:36:00.440 Test: blockdev write read 8 blocks ...passed 00:36:00.440 Test: blockdev write read size > 128k ...passed 00:36:00.440 Test: blockdev write read invalid size ...passed 00:36:00.440 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:36:00.440 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:36:00.440 Test: blockdev write read max offset ...passed 00:36:00.440 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:36:00.440 Test: blockdev writev readv 8 blocks ...passed 00:36:00.440 Test: blockdev writev readv 30 x 1block ...passed 00:36:00.440 Test: blockdev writev readv block ...passed 00:36:00.440 Test: blockdev writev readv size > 128k ...passed 00:36:00.440 Test: blockdev writev readv size > 128k in two iovs ...passed 00:36:00.440 Test: blockdev comparev and writev ...passed 00:36:00.440 Test: blockdev nvme passthru rw ...passed 00:36:00.440 Test: blockdev nvme passthru vendor specific ...passed 00:36:00.440 Test: blockdev nvme admin passthru ...passed 00:36:00.440 Test: blockdev copy ...passed 00:36:00.440 00:36:00.440 Run Summary: Type Total Ran Passed Failed Inactive 00:36:00.440 suites 1 1 n/a 0 0 00:36:00.440 tests 23 23 23 0 0 00:36:00.440 asserts 130 130 130 0 n/a 00:36:00.440 00:36:00.440 Elapsed time = 0.449 seconds 00:36:00.440 0 00:36:00.440 13:56:07 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@294 -- # killprocess 87378 00:36:00.440 13:56:07 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@954 -- # '[' -z 87378 ']' 00:36:00.440 13:56:07 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@958 -- # kill -0 87378 00:36:00.440 13:56:07 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@959 -- # uname 00:36:00.440 13:56:07 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:36:00.440 13:56:07 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 87378 00:36:00.440 13:56:07 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:36:00.440 13:56:07 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:36:00.440 13:56:07 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@972 -- # echo 'killing process with pid 87378' 00:36:00.440 killing process with pid 87378 00:36:00.440 13:56:07 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@973 -- # kill 87378 00:36:00.440 13:56:07 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@978 -- # wait 87378 00:36:01.375 13:56:08 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@295 -- # trap - SIGINT SIGTERM EXIT 00:36:01.375 ************************************ 00:36:01.375 END TEST bdev_bounds 00:36:01.375 ************************************ 00:36:01.375 00:36:01.375 real 0m1.985s 00:36:01.375 user 0m4.912s 00:36:01.375 sys 0m0.292s 00:36:01.375 13:56:08 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@1130 -- # xtrace_disable 00:36:01.375 13:56:08 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:36:01.375 13:56:08 blockdev_raid5f -- bdev/blockdev.sh@799 -- # run_test bdev_nbd nbd_function_test /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json raid5f '' 00:36:01.375 13:56:08 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:36:01.375 13:56:08 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:36:01.375 13:56:08 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:36:01.375 ************************************ 00:36:01.375 START TEST bdev_nbd 00:36:01.375 ************************************ 00:36:01.375 13:56:08 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@1129 -- # nbd_function_test /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json raid5f '' 00:36:01.375 13:56:08 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@299 -- # uname -s 00:36:01.375 13:56:08 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@299 -- # [[ Linux == Linux ]] 00:36:01.375 13:56:08 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@301 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:36:01.375 13:56:08 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@302 -- # local conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:36:01.375 13:56:08 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@303 -- # bdev_all=('raid5f') 00:36:01.375 13:56:08 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@303 -- # local bdev_all 00:36:01.375 13:56:08 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@304 -- # local bdev_num=1 00:36:01.375 13:56:08 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@308 -- # [[ -e /sys/module/nbd ]] 00:36:01.376 13:56:08 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@310 -- # nbd_all=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11' '/dev/nbd12' '/dev/nbd13' '/dev/nbd14' '/dev/nbd15' '/dev/nbd2' '/dev/nbd3' '/dev/nbd4' '/dev/nbd5' '/dev/nbd6' '/dev/nbd7' '/dev/nbd8' '/dev/nbd9') 00:36:01.376 13:56:08 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@310 -- # local nbd_all 00:36:01.376 13:56:08 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@311 -- # bdev_num=1 00:36:01.376 13:56:08 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@313 -- # nbd_list=('/dev/nbd0') 00:36:01.376 13:56:08 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@313 -- # local nbd_list 00:36:01.376 13:56:08 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@314 -- # bdev_list=('raid5f') 00:36:01.376 13:56:08 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@314 -- # local bdev_list 00:36:01.376 13:56:08 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@317 -- # nbd_pid=87431 00:36:01.376 13:56:08 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@318 -- # trap 'cleanup; killprocess $nbd_pid' SIGINT SIGTERM EXIT 00:36:01.376 13:56:08 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@319 -- # waitforlisten 87431 /var/tmp/spdk-nbd.sock 00:36:01.376 13:56:08 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@835 -- # '[' -z 87431 ']' 00:36:01.376 13:56:08 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:36:01.376 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:36:01.376 13:56:08 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@840 -- # local max_retries=100 00:36:01.376 13:56:08 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:36:01.376 13:56:08 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@844 -- # xtrace_disable 00:36:01.376 13:56:08 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:36:01.376 13:56:08 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@316 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-nbd.sock -i 0 --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json '' 00:36:01.376 [2024-11-20 13:56:08.324379] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:36:01.376 [2024-11-20 13:56:08.324478] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:36:01.633 [2024-11-20 13:56:08.479012] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:36:01.633 [2024-11-20 13:56:08.595432] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:36:02.204 13:56:09 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:36:02.204 13:56:09 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@868 -- # return 0 00:36:02.204 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@321 -- # nbd_rpc_start_stop_verify /var/tmp/spdk-nbd.sock raid5f 00:36:02.204 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@113 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:36:02.204 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@114 -- # bdev_list=('raid5f') 00:36:02.204 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@114 -- # local bdev_list 00:36:02.204 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@116 -- # nbd_start_disks_without_nbd_idx /var/tmp/spdk-nbd.sock raid5f 00:36:02.204 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@22 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:36:02.204 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@23 -- # bdev_list=('raid5f') 00:36:02.204 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@23 -- # local bdev_list 00:36:02.204 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@24 -- # local i 00:36:02.204 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@25 -- # local nbd_device 00:36:02.204 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i = 0 )) 00:36:02.204 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 1 )) 00:36:02.204 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid5f 00:36:02.461 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd0 00:36:02.461 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd0 00:36:02.461 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd0 00:36:02.461 13:56:09 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:36:02.461 13:56:09 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@873 -- # local i 00:36:02.461 13:56:09 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:36:02.461 13:56:09 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:36:02.461 13:56:09 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:36:02.461 13:56:09 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@877 -- # break 00:36:02.461 13:56:09 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:36:02.461 13:56:09 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:36:02.461 13:56:09 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:36:02.461 1+0 records in 00:36:02.461 1+0 records out 00:36:02.461 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000286621 s, 14.3 MB/s 00:36:02.461 13:56:09 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:36:02.461 13:56:09 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # size=4096 00:36:02.461 13:56:09 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:36:02.461 13:56:09 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:36:02.461 13:56:09 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@893 -- # return 0 00:36:02.462 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:36:02.462 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 1 )) 00:36:02.462 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@118 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:36:02.720 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@118 -- # nbd_disks_json='[ 00:36:02.720 { 00:36:02.720 "nbd_device": "/dev/nbd0", 00:36:02.720 "bdev_name": "raid5f" 00:36:02.720 } 00:36:02.720 ]' 00:36:02.720 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # nbd_disks_name=($(echo "${nbd_disks_json}" | jq -r '.[] | .nbd_device')) 00:36:02.720 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # echo '[ 00:36:02.720 { 00:36:02.720 "nbd_device": "/dev/nbd0", 00:36:02.720 "bdev_name": "raid5f" 00:36:02.720 } 00:36:02.720 ]' 00:36:02.720 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # jq -r '.[] | .nbd_device' 00:36:02.720 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@120 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:36:02.720 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:36:02.720 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:36:02.720 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:36:02.720 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:36:02.720 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:36:02.720 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:36:02.978 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:36:02.978 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:36:02.978 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:36:02.978 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:36:02.978 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:36:02.978 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:36:02.978 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:36:02.978 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:36:02.978 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@122 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:36:02.978 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:36:02.978 13:56:09 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@122 -- # count=0 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@123 -- # '[' 0 -ne 0 ']' 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@127 -- # return 0 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@322 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock raid5f /dev/nbd0 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@91 -- # bdev_list=('raid5f') 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@91 -- # local bdev_list 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0') 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@92 -- # local nbd_list 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock raid5f /dev/nbd0 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@10 -- # bdev_list=('raid5f') 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@10 -- # local bdev_list 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@11 -- # local nbd_list 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@12 -- # local i 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:36:03.236 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:36:03.237 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid5f /dev/nbd0 00:36:03.494 /dev/nbd0 00:36:03.494 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:36:03.494 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:36:03.494 13:56:10 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:36:03.494 13:56:10 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@873 -- # local i 00:36:03.494 13:56:10 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:36:03.494 13:56:10 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:36:03.494 13:56:10 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:36:03.494 13:56:10 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@877 -- # break 00:36:03.494 13:56:10 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:36:03.494 13:56:10 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:36:03.495 13:56:10 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:36:03.495 1+0 records in 00:36:03.495 1+0 records out 00:36:03.495 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000363082 s, 11.3 MB/s 00:36:03.495 13:56:10 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:36:03.495 13:56:10 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # size=4096 00:36:03.495 13:56:10 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:36:03.495 13:56:10 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:36:03.495 13:56:10 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@893 -- # return 0 00:36:03.495 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:36:03.495 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:36:03.495 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:36:03.495 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:36:03.495 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:36:03.753 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:36:03.753 { 00:36:03.753 "nbd_device": "/dev/nbd0", 00:36:03.753 "bdev_name": "raid5f" 00:36:03.753 } 00:36:03.753 ]' 00:36:03.753 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[ 00:36:03.753 { 00:36:03.753 "nbd_device": "/dev/nbd0", 00:36:03.753 "bdev_name": "raid5f" 00:36:03.753 } 00:36:03.753 ]' 00:36:03.753 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:36:03.753 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:36:03.753 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:36:03.753 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:36:03.753 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=1 00:36:03.753 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 1 00:36:03.753 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@95 -- # count=1 00:36:03.753 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@96 -- # '[' 1 -ne 1 ']' 00:36:03.753 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify /dev/nbd0 write 00:36:03.753 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0') 00:36:03.753 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:36:03.753 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=write 00:36:03.753 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:36:03.753 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:36:03.753 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest bs=4096 count=256 00:36:03.753 256+0 records in 00:36:03.753 256+0 records out 00:36:03.753 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00705661 s, 149 MB/s 00:36:03.753 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:36:03.753 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:36:03.753 256+0 records in 00:36:03.753 256+0 records out 00:36:03.753 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0263485 s, 39.8 MB/s 00:36:03.753 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify /dev/nbd0 verify 00:36:03.753 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0') 00:36:03.753 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:36:03.753 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=verify 00:36:03.753 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:36:03.754 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:36:03.754 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:36:03.754 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:36:03.754 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest /dev/nbd0 00:36:03.754 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:36:03.754 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:36:03.754 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:36:03.754 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:36:03.754 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:36:03.754 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:36:03.754 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:36:03.754 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:36:04.012 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:36:04.012 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:36:04.012 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:36:04.012 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:36:04.012 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:36:04.012 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:36:04.012 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:36:04.012 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:36:04.012 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:36:04.012 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:36:04.012 13:56:10 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:36:04.270 13:56:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:36:04.270 13:56:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:36:04.270 13:56:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:36:04.270 13:56:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:36:04.270 13:56:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:36:04.270 13:56:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:36:04.270 13:56:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:36:04.270 13:56:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:36:04.270 13:56:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:36:04.270 13:56:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@104 -- # count=0 00:36:04.270 13:56:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:36:04.270 13:56:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@109 -- # return 0 00:36:04.270 13:56:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@323 -- # nbd_with_lvol_verify /var/tmp/spdk-nbd.sock /dev/nbd0 00:36:04.270 13:56:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@131 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:36:04.270 13:56:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@132 -- # local nbd=/dev/nbd0 00:36:04.270 13:56:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@134 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create -b malloc_lvol_verify 16 512 00:36:04.529 malloc_lvol_verify 00:36:04.529 13:56:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@135 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create_lvstore malloc_lvol_verify lvs 00:36:04.810 3c2e3819-688f-4508-8f47-6a6bff8d43a9 00:36:04.810 13:56:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@136 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create lvol 4 -l lvs 00:36:04.810 d96429e6-f2cf-41f2-a24e-7d47fa0090cc 00:36:04.810 13:56:11 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@137 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk lvs/lvol /dev/nbd0 00:36:05.068 /dev/nbd0 00:36:05.068 13:56:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@139 -- # wait_for_nbd_set_capacity /dev/nbd0 00:36:05.068 13:56:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@146 -- # local nbd=nbd0 00:36:05.068 13:56:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@148 -- # [[ -e /sys/block/nbd0/size ]] 00:36:05.068 13:56:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@150 -- # (( 8192 == 0 )) 00:36:05.068 13:56:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@141 -- # mkfs.ext4 /dev/nbd0 00:36:05.068 mke2fs 1.47.0 (5-Feb-2023) 00:36:05.068 Discarding device blocks: 0/4096 done 00:36:05.068 Creating filesystem with 4096 1k blocks and 1024 inodes 00:36:05.068 00:36:05.068 Allocating group tables: 0/1 done 00:36:05.068 Writing inode tables: 0/1 done 00:36:05.068 Creating journal (1024 blocks): done 00:36:05.068 Writing superblocks and filesystem accounting information: 0/1 done 00:36:05.068 00:36:05.068 13:56:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@142 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:36:05.068 13:56:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:36:05.068 13:56:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:36:05.068 13:56:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:36:05.068 13:56:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:36:05.068 13:56:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:36:05.068 13:56:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:36:05.327 13:56:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:36:05.327 13:56:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:36:05.327 13:56:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:36:05.327 13:56:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:36:05.327 13:56:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:36:05.327 13:56:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:36:05.327 13:56:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:36:05.327 13:56:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:36:05.327 13:56:12 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@325 -- # killprocess 87431 00:36:05.327 13:56:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@954 -- # '[' -z 87431 ']' 00:36:05.327 13:56:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@958 -- # kill -0 87431 00:36:05.327 13:56:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@959 -- # uname 00:36:05.327 13:56:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:36:05.327 13:56:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 87431 00:36:05.327 13:56:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:36:05.327 killing process with pid 87431 00:36:05.327 13:56:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:36:05.327 13:56:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@972 -- # echo 'killing process with pid 87431' 00:36:05.327 13:56:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@973 -- # kill 87431 00:36:05.327 13:56:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@978 -- # wait 87431 00:36:06.260 13:56:13 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@326 -- # trap - SIGINT SIGTERM EXIT 00:36:06.260 00:36:06.260 real 0m4.942s 00:36:06.260 user 0m7.002s 00:36:06.260 sys 0m1.075s 00:36:06.260 13:56:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@1130 -- # xtrace_disable 00:36:06.260 13:56:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:36:06.260 ************************************ 00:36:06.260 END TEST bdev_nbd 00:36:06.260 ************************************ 00:36:06.260 13:56:13 blockdev_raid5f -- bdev/blockdev.sh@800 -- # [[ y == y ]] 00:36:06.261 13:56:13 blockdev_raid5f -- bdev/blockdev.sh@801 -- # '[' raid5f = nvme ']' 00:36:06.261 13:56:13 blockdev_raid5f -- bdev/blockdev.sh@801 -- # '[' raid5f = gpt ']' 00:36:06.261 13:56:13 blockdev_raid5f -- bdev/blockdev.sh@805 -- # run_test bdev_fio fio_test_suite '' 00:36:06.261 13:56:13 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:36:06.261 13:56:13 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:36:06.261 13:56:13 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:36:06.261 ************************************ 00:36:06.261 START TEST bdev_fio 00:36:06.261 ************************************ 00:36:06.261 13:56:13 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1129 -- # fio_test_suite '' 00:36:06.261 13:56:13 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@330 -- # local env_context 00:36:06.261 13:56:13 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@334 -- # pushd /home/vagrant/spdk_repo/spdk/test/bdev 00:36:06.261 /home/vagrant/spdk_repo/spdk/test/bdev /home/vagrant/spdk_repo/spdk 00:36:06.261 13:56:13 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@335 -- # trap 'rm -f ./*.state; popd; exit 1' SIGINT SIGTERM EXIT 00:36:06.261 13:56:13 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # echo '' 00:36:06.261 13:56:13 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # sed s/--env-context=// 00:36:06.261 13:56:13 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # env_context= 00:36:06.261 13:56:13 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@339 -- # fio_config_gen /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio verify AIO '' 00:36:06.261 13:56:13 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1284 -- # local config_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:36:06.261 13:56:13 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1285 -- # local workload=verify 00:36:06.261 13:56:13 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1286 -- # local bdev_type=AIO 00:36:06.261 13:56:13 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1287 -- # local env_context= 00:36:06.261 13:56:13 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1288 -- # local fio_dir=/usr/src/fio 00:36:06.261 13:56:13 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1290 -- # '[' -e /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio ']' 00:36:06.261 13:56:13 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -z verify ']' 00:36:06.261 13:56:13 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1299 -- # '[' -n '' ']' 00:36:06.261 13:56:13 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1303 -- # touch /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:36:06.261 13:56:13 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1305 -- # cat 00:36:06.261 13:56:13 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1317 -- # '[' verify == verify ']' 00:36:06.261 13:56:13 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1318 -- # cat 00:36:06.261 13:56:13 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1327 -- # '[' AIO == AIO ']' 00:36:06.261 13:56:13 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1328 -- # /usr/src/fio/fio --version 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1328 -- # [[ fio-3.35 == *\f\i\o\-\3* ]] 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1329 -- # echo serialize_overlap=1 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@340 -- # for b in "${bdevs_name[@]}" 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@341 -- # echo '[job_raid5f]' 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@342 -- # echo filename=raid5f 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@346 -- # local 'fio_params=--ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json' 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@348 -- # run_test bdev_fio_rw_verify fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1105 -- # '[' 11 -le 1 ']' 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1111 -- # xtrace_disable 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:36:06.519 ************************************ 00:36:06.519 START TEST bdev_fio_rw_verify 00:36:06.519 ************************************ 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1129 -- # fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1360 -- # fio_plugin /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1341 -- # local fio_dir=/usr/src/fio 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1343 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1343 -- # local sanitizers 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1344 -- # local plugin=/home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # shift 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1347 -- # local asan_lib= 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # grep libasan 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # ldd /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # asan_lib=/usr/lib64/libasan.so.8 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1350 -- # [[ -n /usr/lib64/libasan.so.8 ]] 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1351 -- # break 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1356 -- # LD_PRELOAD='/usr/lib64/libasan.so.8 /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev' 00:36:06.519 13:56:13 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1356 -- # /usr/src/fio/fio --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:36:06.519 job_raid5f: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:36:06.519 fio-3.35 00:36:06.519 Starting 1 thread 00:36:18.730 00:36:18.730 job_raid5f: (groupid=0, jobs=1): err= 0: pid=87623: Wed Nov 20 13:56:24 2024 00:36:18.730 read: IOPS=12.3k, BW=48.1MiB/s (50.5MB/s)(481MiB/10001msec) 00:36:18.730 slat (nsec): min=17962, max=75122, avg=20114.75, stdev=2440.96 00:36:18.730 clat (usec): min=8, max=336, avg=132.78, stdev=49.43 00:36:18.730 lat (usec): min=27, max=360, avg=152.90, stdev=50.08 00:36:18.730 clat percentiles (usec): 00:36:18.730 | 50.000th=[ 135], 99.000th=[ 249], 99.900th=[ 262], 99.990th=[ 285], 00:36:18.730 | 99.999th=[ 310] 00:36:18.730 write: IOPS=12.9k, BW=50.3MiB/s (52.8MB/s)(497MiB/9874msec); 0 zone resets 00:36:18.730 slat (usec): min=7, max=351, avg=16.21, stdev= 2.86 00:36:18.730 clat (usec): min=53, max=1339, avg=296.08, stdev=47.28 00:36:18.730 lat (usec): min=68, max=1506, avg=312.30, stdev=48.73 00:36:18.730 clat percentiles (usec): 00:36:18.730 | 50.000th=[ 297], 99.000th=[ 416], 99.900th=[ 441], 99.990th=[ 1106], 00:36:18.730 | 99.999th=[ 1270] 00:36:18.730 bw ( KiB/s): min=39432, max=56640, per=98.76%, avg=50906.11, stdev=5196.43, samples=19 00:36:18.730 iops : min= 9858, max=14160, avg=12726.53, stdev=1299.11, samples=19 00:36:18.730 lat (usec) : 10=0.01%, 20=0.01%, 50=0.01%, 100=16.54%, 250=39.96% 00:36:18.730 lat (usec) : 500=43.48%, 750=0.01%, 1000=0.01% 00:36:18.730 lat (msec) : 2=0.01% 00:36:18.730 cpu : usr=99.22%, sys=0.24%, ctx=22, majf=0, minf=10073 00:36:18.730 IO depths : 1=7.6%, 2=19.7%, 4=55.3%, 8=17.3%, 16=0.0%, 32=0.0%, >=64=0.0% 00:36:18.730 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:18.730 complete : 0=0.0%, 4=90.0%, 8=10.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:36:18.730 issued rwts: total=123236,127238,0,0 short=0,0,0,0 dropped=0,0,0,0 00:36:18.730 latency : target=0, window=0, percentile=100.00%, depth=8 00:36:18.730 00:36:18.730 Run status group 0 (all jobs): 00:36:18.730 READ: bw=48.1MiB/s (50.5MB/s), 48.1MiB/s-48.1MiB/s (50.5MB/s-50.5MB/s), io=481MiB (505MB), run=10001-10001msec 00:36:18.730 WRITE: bw=50.3MiB/s (52.8MB/s), 50.3MiB/s-50.3MiB/s (52.8MB/s-52.8MB/s), io=497MiB (521MB), run=9874-9874msec 00:36:18.730 ----------------------------------------------------- 00:36:18.730 Suppressions used: 00:36:18.730 count bytes template 00:36:18.731 1 7 /usr/src/fio/parse.c 00:36:18.731 185 17760 /usr/src/fio/iolog.c 00:36:18.731 1 8 libtcmalloc_minimal.so 00:36:18.731 1 904 libcrypto.so 00:36:18.731 ----------------------------------------------------- 00:36:18.731 00:36:18.731 00:36:18.731 real 0m11.928s 00:36:18.731 user 0m12.505s 00:36:18.731 sys 0m0.570s 00:36:18.731 ************************************ 00:36:18.731 END TEST bdev_fio_rw_verify 00:36:18.731 ************************************ 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1130 -- # xtrace_disable 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@10 -- # set +x 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@349 -- # rm -f 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@350 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@353 -- # fio_config_gen /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio trim '' '' 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1284 -- # local config_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1285 -- # local workload=trim 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1286 -- # local bdev_type= 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1287 -- # local env_context= 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1288 -- # local fio_dir=/usr/src/fio 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1290 -- # '[' -e /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio ']' 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -z trim ']' 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1299 -- # '[' -n '' ']' 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1303 -- # touch /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1305 -- # cat 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1317 -- # '[' trim == verify ']' 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1332 -- # '[' trim == trim ']' 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1333 -- # echo rw=trimwrite 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # printf '%s\n' '{' ' "name": "raid5f",' ' "aliases": [' ' "1143f858-e763-4400-8e6a-dd7ee4b79357"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "1143f858-e763-4400-8e6a-dd7ee4b79357",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "raid": {' ' "uuid": "1143f858-e763-4400-8e6a-dd7ee4b79357",' ' "strip_size_kb": 2,' ' "state": "online",' ' "raid_level": "raid5f",' ' "superblock": false,' ' "num_base_bdevs": 3,' ' "num_base_bdevs_discovered": 3,' ' "num_base_bdevs_operational": 3,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc0",' ' "uuid": "2be9841d-5c24-445e-9e69-8fd81a7b244f",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc1",' ' "uuid": "0ea55a6d-7d4c-47c3-97ad-326b81a47106",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc2",' ' "uuid": "1e0b630f-0da7-4ff0-a129-ea28b016558c",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # jq -r 'select(.supported_io_types.unmap == true) | .name' 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # [[ -n '' ]] 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@360 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@361 -- # popd 00:36:18.731 /home/vagrant/spdk_repo/spdk 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@362 -- # trap - SIGINT SIGTERM EXIT 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@363 -- # return 0 00:36:18.731 00:36:18.731 real 0m12.112s 00:36:18.731 user 0m12.576s 00:36:18.731 sys 0m0.650s 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1130 -- # xtrace_disable 00:36:18.731 ************************************ 00:36:18.731 END TEST bdev_fio 00:36:18.731 ************************************ 00:36:18.731 13:56:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:36:18.731 13:56:25 blockdev_raid5f -- bdev/blockdev.sh@812 -- # trap cleanup SIGINT SIGTERM EXIT 00:36:18.731 13:56:25 blockdev_raid5f -- bdev/blockdev.sh@814 -- # run_test bdev_verify /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:36:18.731 13:56:25 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 16 -le 1 ']' 00:36:18.731 13:56:25 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:36:18.731 13:56:25 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:36:18.731 ************************************ 00:36:18.731 START TEST bdev_verify 00:36:18.731 ************************************ 00:36:18.731 13:56:25 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:36:18.731 [2024-11-20 13:56:25.487090] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:36:18.731 [2024-11-20 13:56:25.487221] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87781 ] 00:36:18.731 [2024-11-20 13:56:25.644107] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:36:18.731 [2024-11-20 13:56:25.750376] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:36:18.731 [2024-11-20 13:56:25.750467] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:36:19.297 Running I/O for 5 seconds... 00:36:21.167 17842.00 IOPS, 69.70 MiB/s [2024-11-20T13:56:29.185Z] 18078.00 IOPS, 70.62 MiB/s [2024-11-20T13:56:30.558Z] 19171.00 IOPS, 74.89 MiB/s [2024-11-20T13:56:31.489Z] 19828.50 IOPS, 77.46 MiB/s [2024-11-20T13:56:31.489Z] 20022.80 IOPS, 78.21 MiB/s 00:36:24.430 Latency(us) 00:36:24.430 [2024-11-20T13:56:31.489Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:36:24.430 Job: raid5f (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:36:24.430 Verification LBA range: start 0x0 length 0x2000 00:36:24.430 raid5f : 5.01 10332.61 40.36 0.00 0.00 18542.12 165.42 15829.46 00:36:24.430 Job: raid5f (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:36:24.430 Verification LBA range: start 0x2000 length 0x2000 00:36:24.430 raid5f : 5.01 9691.98 37.86 0.00 0.00 19744.33 78.38 18047.61 00:36:24.430 [2024-11-20T13:56:31.489Z] =================================================================================================================== 00:36:24.430 [2024-11-20T13:56:31.489Z] Total : 20024.59 78.22 0.00 0.00 19124.22 78.38 18047.61 00:36:25.000 00:36:25.000 real 0m6.452s 00:36:25.000 user 0m12.008s 00:36:25.000 sys 0m0.230s 00:36:25.000 13:56:31 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@1130 -- # xtrace_disable 00:36:25.000 13:56:31 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@10 -- # set +x 00:36:25.000 ************************************ 00:36:25.000 END TEST bdev_verify 00:36:25.000 ************************************ 00:36:25.000 13:56:31 blockdev_raid5f -- bdev/blockdev.sh@815 -- # run_test bdev_verify_big_io /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:36:25.000 13:56:31 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 16 -le 1 ']' 00:36:25.000 13:56:31 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:36:25.000 13:56:31 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:36:25.000 ************************************ 00:36:25.000 START TEST bdev_verify_big_io 00:36:25.000 ************************************ 00:36:25.000 13:56:31 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:36:25.000 [2024-11-20 13:56:31.982221] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:36:25.000 [2024-11-20 13:56:31.982442] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87868 ] 00:36:25.259 [2024-11-20 13:56:32.128203] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:36:25.259 [2024-11-20 13:56:32.228520] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:36:25.259 [2024-11-20 13:56:32.228606] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:36:25.826 Running I/O for 5 seconds... 00:36:27.696 1014.00 IOPS, 63.38 MiB/s [2024-11-20T13:56:35.690Z] 1141.00 IOPS, 71.31 MiB/s [2024-11-20T13:56:37.064Z] 1162.33 IOPS, 72.65 MiB/s [2024-11-20T13:56:37.998Z] 1189.25 IOPS, 74.33 MiB/s [2024-11-20T13:56:37.998Z] 1218.40 IOPS, 76.15 MiB/s 00:36:30.939 Latency(us) 00:36:30.939 [2024-11-20T13:56:37.998Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:36:30.939 Job: raid5f (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:36:30.939 Verification LBA range: start 0x0 length 0x200 00:36:30.939 raid5f : 5.15 616.12 38.51 0.00 0.00 5059521.20 123.67 216167.98 00:36:30.939 Job: raid5f (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:36:30.939 Verification LBA range: start 0x200 length 0x200 00:36:30.939 raid5f : 5.20 610.21 38.14 0.00 0.00 5209272.64 124.46 219394.36 00:36:30.939 [2024-11-20T13:56:37.998Z] =================================================================================================================== 00:36:30.939 [2024-11-20T13:56:37.998Z] Total : 1226.33 76.65 0.00 0.00 5134373.33 123.67 219394.36 00:36:31.524 ************************************ 00:36:31.524 END TEST bdev_verify_big_io 00:36:31.524 ************************************ 00:36:31.524 00:36:31.524 real 0m6.628s 00:36:31.524 user 0m12.415s 00:36:31.524 sys 0m0.201s 00:36:31.524 13:56:38 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:36:31.524 13:56:38 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@10 -- # set +x 00:36:31.783 13:56:38 blockdev_raid5f -- bdev/blockdev.sh@816 -- # run_test bdev_write_zeroes /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:36:31.783 13:56:38 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 13 -le 1 ']' 00:36:31.783 13:56:38 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:36:31.783 13:56:38 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:36:31.783 ************************************ 00:36:31.783 START TEST bdev_write_zeroes 00:36:31.783 ************************************ 00:36:31.783 13:56:38 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:36:31.783 [2024-11-20 13:56:38.672367] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:36:31.783 [2024-11-20 13:56:38.672628] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87961 ] 00:36:31.783 [2024-11-20 13:56:38.830252] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:36:32.041 [2024-11-20 13:56:38.919351] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:36:32.298 Running I/O for 1 seconds... 00:36:33.231 29559.00 IOPS, 115.46 MiB/s 00:36:33.231 Latency(us) 00:36:33.231 [2024-11-20T13:56:40.290Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:36:33.231 Job: raid5f (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:36:33.231 raid5f : 1.01 29521.88 115.32 0.00 0.00 4322.55 1241.40 6099.89 00:36:33.231 [2024-11-20T13:56:40.290Z] =================================================================================================================== 00:36:33.231 [2024-11-20T13:56:40.290Z] Total : 29521.88 115.32 0.00 0.00 4322.55 1241.40 6099.89 00:36:34.166 ************************************ 00:36:34.166 END TEST bdev_write_zeroes 00:36:34.166 ************************************ 00:36:34.166 00:36:34.166 real 0m2.420s 00:36:34.166 user 0m2.090s 00:36:34.166 sys 0m0.209s 00:36:34.166 13:56:41 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@1130 -- # xtrace_disable 00:36:34.166 13:56:41 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@10 -- # set +x 00:36:34.166 13:56:41 blockdev_raid5f -- bdev/blockdev.sh@819 -- # run_test bdev_json_nonenclosed /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:36:34.166 13:56:41 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 13 -le 1 ']' 00:36:34.166 13:56:41 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:36:34.166 13:56:41 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:36:34.166 ************************************ 00:36:34.166 START TEST bdev_json_nonenclosed 00:36:34.166 ************************************ 00:36:34.166 13:56:41 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:36:34.166 [2024-11-20 13:56:41.153466] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:36:34.166 [2024-11-20 13:56:41.153719] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid88003 ] 00:36:34.427 [2024-11-20 13:56:41.311271] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:36:34.427 [2024-11-20 13:56:41.401740] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:36:34.427 [2024-11-20 13:56:41.401813] json_config.c: 608:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: not enclosed in {}. 00:36:34.427 [2024-11-20 13:56:41.401833] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:36:34.427 [2024-11-20 13:56:41.401841] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:36:34.687 ************************************ 00:36:34.687 END TEST bdev_json_nonenclosed 00:36:34.687 ************************************ 00:36:34.687 00:36:34.687 real 0m0.463s 00:36:34.687 user 0m0.259s 00:36:34.687 sys 0m0.100s 00:36:34.687 13:56:41 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@1130 -- # xtrace_disable 00:36:34.687 13:56:41 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@10 -- # set +x 00:36:34.687 13:56:41 blockdev_raid5f -- bdev/blockdev.sh@822 -- # run_test bdev_json_nonarray /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:36:34.687 13:56:41 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 13 -le 1 ']' 00:36:34.687 13:56:41 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:36:34.687 13:56:41 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:36:34.687 ************************************ 00:36:34.687 START TEST bdev_json_nonarray 00:36:34.687 ************************************ 00:36:34.687 13:56:41 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:36:34.687 [2024-11-20 13:56:41.666819] Starting SPDK v25.01-pre git sha1 a361eb5e2 / DPDK 24.03.0 initialization... 00:36:34.687 [2024-11-20 13:56:41.666937] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid88029 ] 00:36:34.947 [2024-11-20 13:56:41.823204] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:36:34.947 [2024-11-20 13:56:41.914647] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:36:34.947 [2024-11-20 13:56:41.914887] json_config.c: 614:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: 'subsystems' should be an array. 00:36:34.947 [2024-11-20 13:56:41.914909] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:36:34.947 [2024-11-20 13:56:41.914924] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:36:35.207 ************************************ 00:36:35.207 END TEST bdev_json_nonarray 00:36:35.207 ************************************ 00:36:35.207 00:36:35.207 real 0m0.449s 00:36:35.207 user 0m0.253s 00:36:35.207 sys 0m0.093s 00:36:35.207 13:56:42 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@1130 -- # xtrace_disable 00:36:35.207 13:56:42 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@10 -- # set +x 00:36:35.207 13:56:42 blockdev_raid5f -- bdev/blockdev.sh@824 -- # [[ raid5f == bdev ]] 00:36:35.207 13:56:42 blockdev_raid5f -- bdev/blockdev.sh@832 -- # [[ raid5f == gpt ]] 00:36:35.207 13:56:42 blockdev_raid5f -- bdev/blockdev.sh@836 -- # [[ raid5f == crypto_sw ]] 00:36:35.207 13:56:42 blockdev_raid5f -- bdev/blockdev.sh@848 -- # trap - SIGINT SIGTERM EXIT 00:36:35.207 13:56:42 blockdev_raid5f -- bdev/blockdev.sh@849 -- # cleanup 00:36:35.207 13:56:42 blockdev_raid5f -- bdev/blockdev.sh@23 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/aiofile 00:36:35.207 13:56:42 blockdev_raid5f -- bdev/blockdev.sh@24 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:36:35.207 13:56:42 blockdev_raid5f -- bdev/blockdev.sh@26 -- # [[ raid5f == rbd ]] 00:36:35.207 13:56:42 blockdev_raid5f -- bdev/blockdev.sh@30 -- # [[ raid5f == daos ]] 00:36:35.207 13:56:42 blockdev_raid5f -- bdev/blockdev.sh@34 -- # [[ raid5f = \g\p\t ]] 00:36:35.207 13:56:42 blockdev_raid5f -- bdev/blockdev.sh@40 -- # [[ raid5f == xnvme ]] 00:36:35.207 00:36:35.207 real 0m40.029s 00:36:35.207 user 0m55.353s 00:36:35.207 sys 0m3.811s 00:36:35.207 ************************************ 00:36:35.207 END TEST blockdev_raid5f 00:36:35.207 ************************************ 00:36:35.207 13:56:42 blockdev_raid5f -- common/autotest_common.sh@1130 -- # xtrace_disable 00:36:35.207 13:56:42 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:36:35.207 13:56:42 -- spdk/autotest.sh@194 -- # uname -s 00:36:35.208 13:56:42 -- spdk/autotest.sh@194 -- # [[ Linux == Linux ]] 00:36:35.208 13:56:42 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:36:35.208 13:56:42 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:36:35.208 13:56:42 -- spdk/autotest.sh@207 -- # '[' 0 -eq 1 ']' 00:36:35.208 13:56:42 -- spdk/autotest.sh@256 -- # '[' 0 -eq 1 ']' 00:36:35.208 13:56:42 -- spdk/autotest.sh@260 -- # timing_exit lib 00:36:35.208 13:56:42 -- common/autotest_common.sh@732 -- # xtrace_disable 00:36:35.208 13:56:42 -- common/autotest_common.sh@10 -- # set +x 00:36:35.208 13:56:42 -- spdk/autotest.sh@262 -- # '[' 0 -eq 1 ']' 00:36:35.208 13:56:42 -- spdk/autotest.sh@267 -- # '[' 0 -eq 1 ']' 00:36:35.208 13:56:42 -- spdk/autotest.sh@276 -- # '[' 0 -eq 1 ']' 00:36:35.208 13:56:42 -- spdk/autotest.sh@311 -- # '[' 0 -eq 1 ']' 00:36:35.208 13:56:42 -- spdk/autotest.sh@315 -- # '[' 0 -eq 1 ']' 00:36:35.208 13:56:42 -- spdk/autotest.sh@319 -- # '[' 0 -eq 1 ']' 00:36:35.208 13:56:42 -- spdk/autotest.sh@324 -- # '[' 0 -eq 1 ']' 00:36:35.208 13:56:42 -- spdk/autotest.sh@333 -- # '[' 0 -eq 1 ']' 00:36:35.208 13:56:42 -- spdk/autotest.sh@338 -- # '[' 0 -eq 1 ']' 00:36:35.208 13:56:42 -- spdk/autotest.sh@342 -- # '[' 0 -eq 1 ']' 00:36:35.208 13:56:42 -- spdk/autotest.sh@346 -- # '[' 0 -eq 1 ']' 00:36:35.208 13:56:42 -- spdk/autotest.sh@350 -- # '[' 0 -eq 1 ']' 00:36:35.208 13:56:42 -- spdk/autotest.sh@355 -- # '[' 0 -eq 1 ']' 00:36:35.208 13:56:42 -- spdk/autotest.sh@359 -- # '[' 0 -eq 1 ']' 00:36:35.208 13:56:42 -- spdk/autotest.sh@366 -- # [[ 0 -eq 1 ]] 00:36:35.208 13:56:42 -- spdk/autotest.sh@370 -- # [[ 0 -eq 1 ]] 00:36:35.208 13:56:42 -- spdk/autotest.sh@374 -- # [[ 0 -eq 1 ]] 00:36:35.208 13:56:42 -- spdk/autotest.sh@378 -- # [[ '' -eq 1 ]] 00:36:35.208 13:56:42 -- spdk/autotest.sh@385 -- # trap - SIGINT SIGTERM EXIT 00:36:35.208 13:56:42 -- spdk/autotest.sh@387 -- # timing_enter post_cleanup 00:36:35.208 13:56:42 -- common/autotest_common.sh@726 -- # xtrace_disable 00:36:35.208 13:56:42 -- common/autotest_common.sh@10 -- # set +x 00:36:35.208 13:56:42 -- spdk/autotest.sh@388 -- # autotest_cleanup 00:36:35.208 13:56:42 -- common/autotest_common.sh@1396 -- # local autotest_es=0 00:36:35.208 13:56:42 -- common/autotest_common.sh@1397 -- # xtrace_disable 00:36:35.208 13:56:42 -- common/autotest_common.sh@10 -- # set +x 00:36:36.589 INFO: APP EXITING 00:36:36.590 INFO: killing all VMs 00:36:36.590 INFO: killing vhost app 00:36:36.590 INFO: EXIT DONE 00:36:36.848 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:36:36.848 Waiting for block devices as requested 00:36:36.848 0000:00:11.0 (1b36 0010): uio_pci_generic -> nvme 00:36:36.848 0000:00:10.0 (1b36 0010): uio_pci_generic -> nvme 00:36:37.791 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:36:37.791 Cleaning 00:36:37.791 Removing: /var/run/dpdk/spdk0/config 00:36:37.791 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-0 00:36:37.791 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-1 00:36:37.791 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-2 00:36:37.791 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-3 00:36:37.791 Removing: /var/run/dpdk/spdk0/fbarray_memzone 00:36:37.791 Removing: /var/run/dpdk/spdk0/hugepage_info 00:36:37.791 Removing: /dev/shm/spdk_tgt_trace.pid56103 00:36:37.791 Removing: /var/run/dpdk/spdk0 00:36:37.791 Removing: /var/run/dpdk/spdk_pid55912 00:36:37.791 Removing: /var/run/dpdk/spdk_pid56103 00:36:37.791 Removing: /var/run/dpdk/spdk_pid56321 00:36:37.791 Removing: /var/run/dpdk/spdk_pid56414 00:36:37.791 Removing: /var/run/dpdk/spdk_pid56454 00:36:37.791 Removing: /var/run/dpdk/spdk_pid56576 00:36:37.791 Removing: /var/run/dpdk/spdk_pid56594 00:36:37.791 Removing: /var/run/dpdk/spdk_pid56793 00:36:37.791 Removing: /var/run/dpdk/spdk_pid56886 00:36:37.791 Removing: /var/run/dpdk/spdk_pid56983 00:36:37.791 Removing: /var/run/dpdk/spdk_pid57093 00:36:37.791 Removing: /var/run/dpdk/spdk_pid57190 00:36:37.791 Removing: /var/run/dpdk/spdk_pid57230 00:36:37.791 Removing: /var/run/dpdk/spdk_pid57266 00:36:37.791 Removing: /var/run/dpdk/spdk_pid57337 00:36:37.791 Removing: /var/run/dpdk/spdk_pid57421 00:36:37.791 Removing: /var/run/dpdk/spdk_pid57857 00:36:37.791 Removing: /var/run/dpdk/spdk_pid57910 00:36:37.791 Removing: /var/run/dpdk/spdk_pid57973 00:36:37.791 Removing: /var/run/dpdk/spdk_pid57989 00:36:37.791 Removing: /var/run/dpdk/spdk_pid58091 00:36:37.791 Removing: /var/run/dpdk/spdk_pid58107 00:36:37.791 Removing: /var/run/dpdk/spdk_pid58204 00:36:37.791 Removing: /var/run/dpdk/spdk_pid58219 00:36:37.791 Removing: /var/run/dpdk/spdk_pid58278 00:36:37.791 Removing: /var/run/dpdk/spdk_pid58296 00:36:37.791 Removing: /var/run/dpdk/spdk_pid58349 00:36:37.791 Removing: /var/run/dpdk/spdk_pid58367 00:36:37.791 Removing: /var/run/dpdk/spdk_pid58527 00:36:37.791 Removing: /var/run/dpdk/spdk_pid58564 00:36:37.791 Removing: /var/run/dpdk/spdk_pid58647 00:36:37.791 Removing: /var/run/dpdk/spdk_pid59879 00:36:37.791 Removing: /var/run/dpdk/spdk_pid60085 00:36:37.791 Removing: /var/run/dpdk/spdk_pid60216 00:36:37.791 Removing: /var/run/dpdk/spdk_pid60829 00:36:37.791 Removing: /var/run/dpdk/spdk_pid61030 00:36:37.791 Removing: /var/run/dpdk/spdk_pid61159 00:36:37.791 Removing: /var/run/dpdk/spdk_pid61769 00:36:37.791 Removing: /var/run/dpdk/spdk_pid62077 00:36:37.791 Removing: /var/run/dpdk/spdk_pid62206 00:36:37.791 Removing: /var/run/dpdk/spdk_pid63533 00:36:37.791 Removing: /var/run/dpdk/spdk_pid63770 00:36:37.791 Removing: /var/run/dpdk/spdk_pid63904 00:36:37.791 Removing: /var/run/dpdk/spdk_pid65223 00:36:37.791 Removing: /var/run/dpdk/spdk_pid65464 00:36:37.791 Removing: /var/run/dpdk/spdk_pid65600 00:36:37.791 Removing: /var/run/dpdk/spdk_pid66919 00:36:37.791 Removing: /var/run/dpdk/spdk_pid67342 00:36:37.791 Removing: /var/run/dpdk/spdk_pid67477 00:36:37.791 Removing: /var/run/dpdk/spdk_pid68890 00:36:37.791 Removing: /var/run/dpdk/spdk_pid69138 00:36:37.791 Removing: /var/run/dpdk/spdk_pid69273 00:36:37.791 Removing: /var/run/dpdk/spdk_pid70681 00:36:37.791 Removing: /var/run/dpdk/spdk_pid70929 00:36:37.791 Removing: /var/run/dpdk/spdk_pid71058 00:36:37.791 Removing: /var/run/dpdk/spdk_pid72470 00:36:37.791 Removing: /var/run/dpdk/spdk_pid72932 00:36:37.791 Removing: /var/run/dpdk/spdk_pid73072 00:36:37.791 Removing: /var/run/dpdk/spdk_pid73199 00:36:37.791 Removing: /var/run/dpdk/spdk_pid73609 00:36:37.791 Removing: /var/run/dpdk/spdk_pid74319 00:36:37.791 Removing: /var/run/dpdk/spdk_pid74699 00:36:37.791 Removing: /var/run/dpdk/spdk_pid75361 00:36:37.791 Removing: /var/run/dpdk/spdk_pid75797 00:36:37.791 Removing: /var/run/dpdk/spdk_pid76532 00:36:37.791 Removing: /var/run/dpdk/spdk_pid76938 00:36:37.791 Removing: /var/run/dpdk/spdk_pid78819 00:36:37.791 Removing: /var/run/dpdk/spdk_pid79242 00:36:37.791 Removing: /var/run/dpdk/spdk_pid79660 00:36:37.791 Removing: /var/run/dpdk/spdk_pid81656 00:36:37.791 Removing: /var/run/dpdk/spdk_pid82119 00:36:37.791 Removing: /var/run/dpdk/spdk_pid82619 00:36:37.791 Removing: /var/run/dpdk/spdk_pid83649 00:36:37.791 Removing: /var/run/dpdk/spdk_pid83961 00:36:37.791 Removing: /var/run/dpdk/spdk_pid84859 00:36:37.791 Removing: /var/run/dpdk/spdk_pid85165 00:36:37.791 Removing: /var/run/dpdk/spdk_pid86072 00:36:37.791 Removing: /var/run/dpdk/spdk_pid86378 00:36:37.791 Removing: /var/run/dpdk/spdk_pid87032 00:36:37.791 Removing: /var/run/dpdk/spdk_pid87291 00:36:37.791 Removing: /var/run/dpdk/spdk_pid87336 00:36:37.791 Removing: /var/run/dpdk/spdk_pid87378 00:36:37.791 Removing: /var/run/dpdk/spdk_pid87608 00:36:37.791 Removing: /var/run/dpdk/spdk_pid87781 00:36:37.791 Removing: /var/run/dpdk/spdk_pid87868 00:36:37.791 Removing: /var/run/dpdk/spdk_pid87961 00:36:37.791 Removing: /var/run/dpdk/spdk_pid88003 00:36:37.791 Removing: /var/run/dpdk/spdk_pid88029 00:36:37.791 Clean 00:36:38.052 13:56:44 -- common/autotest_common.sh@1453 -- # return 0 00:36:38.052 13:56:44 -- spdk/autotest.sh@389 -- # timing_exit post_cleanup 00:36:38.052 13:56:44 -- common/autotest_common.sh@732 -- # xtrace_disable 00:36:38.052 13:56:44 -- common/autotest_common.sh@10 -- # set +x 00:36:38.052 13:56:44 -- spdk/autotest.sh@391 -- # timing_exit autotest 00:36:38.052 13:56:44 -- common/autotest_common.sh@732 -- # xtrace_disable 00:36:38.052 13:56:44 -- common/autotest_common.sh@10 -- # set +x 00:36:38.052 13:56:44 -- spdk/autotest.sh@392 -- # chmod a+r /home/vagrant/spdk_repo/spdk/../output/timing.txt 00:36:38.052 13:56:44 -- spdk/autotest.sh@394 -- # [[ -f /home/vagrant/spdk_repo/spdk/../output/udev.log ]] 00:36:38.052 13:56:44 -- spdk/autotest.sh@394 -- # rm -f /home/vagrant/spdk_repo/spdk/../output/udev.log 00:36:38.052 13:56:44 -- spdk/autotest.sh@396 -- # [[ y == y ]] 00:36:38.052 13:56:44 -- spdk/autotest.sh@398 -- # hostname 00:36:38.052 13:56:44 -- spdk/autotest.sh@398 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -d /home/vagrant/spdk_repo/spdk -t fedora39-cloud-1721788873-2326 -o /home/vagrant/spdk_repo/spdk/../output/cov_test.info 00:36:38.314 geninfo: WARNING: invalid characters removed from testname! 00:37:04.889 13:57:08 -- spdk/autotest.sh@399 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -a /home/vagrant/spdk_repo/spdk/../output/cov_base.info -a /home/vagrant/spdk_repo/spdk/../output/cov_test.info -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:37:04.889 13:57:11 -- spdk/autotest.sh@400 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/dpdk/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:37:06.262 13:57:12 -- spdk/autotest.sh@404 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info --ignore-errors unused,unused '/usr/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:37:08.163 13:57:15 -- spdk/autotest.sh@405 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/examples/vmd/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:37:10.071 13:57:17 -- spdk/autotest.sh@406 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/app/spdk_lspci/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:37:11.971 13:57:18 -- spdk/autotest.sh@407 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/app/spdk_top/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:37:13.873 13:57:20 -- spdk/autotest.sh@408 -- # rm -f cov_base.info cov_test.info OLD_STDOUT OLD_STDERR 00:37:13.873 13:57:20 -- spdk/autorun.sh@1 -- $ timing_finish 00:37:13.873 13:57:20 -- common/autotest_common.sh@738 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/timing.txt ]] 00:37:13.873 13:57:20 -- common/autotest_common.sh@740 -- $ flamegraph=/usr/local/FlameGraph/flamegraph.pl 00:37:13.873 13:57:20 -- common/autotest_common.sh@741 -- $ [[ -x /usr/local/FlameGraph/flamegraph.pl ]] 00:37:13.873 13:57:20 -- common/autotest_common.sh@744 -- $ /usr/local/FlameGraph/flamegraph.pl --title 'Build Timing' --nametype Step: --countname seconds /home/vagrant/spdk_repo/spdk/../output/timing.txt 00:37:13.873 + [[ -n 4990 ]] 00:37:13.873 + sudo kill 4990 00:37:13.884 [Pipeline] } 00:37:13.904 [Pipeline] // timeout 00:37:13.908 [Pipeline] } 00:37:13.924 [Pipeline] // stage 00:37:13.928 [Pipeline] } 00:37:13.945 [Pipeline] // catchError 00:37:13.956 [Pipeline] stage 00:37:13.958 [Pipeline] { (Stop VM) 00:37:13.971 [Pipeline] sh 00:37:14.255 + vagrant halt 00:37:16.797 ==> default: Halting domain... 00:37:20.096 [Pipeline] sh 00:37:20.374 + vagrant destroy -f 00:37:22.904 ==> default: Removing domain... 00:37:22.971 [Pipeline] sh 00:37:23.251 + mv output /var/jenkins/workspace/raid-vg-autotest/output 00:37:23.258 [Pipeline] } 00:37:23.274 [Pipeline] // stage 00:37:23.279 [Pipeline] } 00:37:23.292 [Pipeline] // dir 00:37:23.297 [Pipeline] } 00:37:23.310 [Pipeline] // wrap 00:37:23.316 [Pipeline] } 00:37:23.328 [Pipeline] // catchError 00:37:23.335 [Pipeline] stage 00:37:23.337 [Pipeline] { (Epilogue) 00:37:23.348 [Pipeline] sh 00:37:23.625 + jbp/jenkins/jjb-config/jobs/scripts/compress_artifacts.sh 00:37:28.904 [Pipeline] catchError 00:37:28.906 [Pipeline] { 00:37:28.920 [Pipeline] sh 00:37:29.201 + jbp/jenkins/jjb-config/jobs/scripts/check_artifacts_size.sh 00:37:29.202 Artifacts sizes are good 00:37:29.229 [Pipeline] } 00:37:29.237 [Pipeline] // catchError 00:37:29.245 [Pipeline] archiveArtifacts 00:37:29.249 Archiving artifacts 00:37:29.330 [Pipeline] cleanWs 00:37:29.342 [WS-CLEANUP] Deleting project workspace... 00:37:29.342 [WS-CLEANUP] Deferred wipeout is used... 00:37:29.350 [WS-CLEANUP] done 00:37:29.352 [Pipeline] } 00:37:29.364 [Pipeline] // stage 00:37:29.369 [Pipeline] } 00:37:29.379 [Pipeline] // node 00:37:29.383 [Pipeline] End of Pipeline 00:37:29.416 Finished: SUCCESS